For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | SimFG's commentsregister

Thanks Shivansh for his article


Lior's tweet: https://twitter.com/AlphaSignalAI/status/1677348799801425920

You can cut your GPT API expenses by 50% through caching using LangChain and GPTCache.

You will also benefit from significant response time increase and API rate limit restrictions.


Maybe [GPTCache](https://github.com/zilliztech/GPTCache) can make it more attractive, because similar problems can be less expensive, and can also be responded to faster. Of course, the specific configuration needs to be based on real usage scenarios.


At present, we have started applying GPT cache to online products. According to existing data, the effect is good. In fact, it can also be found that some LLM products actually have caching capabilities, such as Github Copilot. However, the cache required by different products varies, and the caching mechanism is worth noting.


Maybe users can discuss about llm output? The question asker can have LLM reproduce the results based on the discussion.

Of course, there are many user-related issues involved.


People are already angry that LLMs ingested their knowledge and will use that knowledge to replace them. They're not going to do this without being paid.


Cool, I prefer the OpenAI-Compatible api. Although this is not very technically difficult, it is really intimate, because it make me feel free to use all ChatGPT applications.


Thanks! Please try it out and share any feedback you might have.


Looks great! I'm planning to integrate it into my new project(to-chatgpt: https://github.com/SimFG/to-chatgpt), which will provide users of the ChatGPT applications with a wider range of LLM service options.


Looking forward to it!

OpenLLM is adding a OpenAI-compatible API layer, which will make it even easier to migrate LLM apps built around OpenAI's API spec. Feel free to join our Discord community and discuss more!


GPT3.5 answer: If you have limited budget or machine resources, there are several ways to obtain a good large-scale model experience:

1. Pretrained Models: You can use pre-trained models that have already been trained on large datasets. These models are available for free and can be fine-tuned on your specific task with limited computational resources.

2. Transfer Learning: Transfer learning can be used to leverage the knowledge of pre-trained models for your specific task. You can use a pre-trained model as a starting point and fine-tune it on your data to obtain good results.

3. Data Augmentation: Data augmentation techniques can be used to increase the size of your training data. This can help in improving the performance of your model without requiring additional computational resources.

4. Model Compression: Model compression techniques can be used to reduce the size of your model without losing much accuracy. This can help in improving the efficiency of your model and reducing computational resources.

5. Cloud Computing: Cloud computing services like AWS, Google Cloud, and Microsoft Azure provide access to high-performance computing resources at affordable prices. You can use these services to train your models on large datasets with limited computational resources.

Overall, with careful planning and thoughtful use of available resources, it is possible to obtain a good large-scale model experience even with limited budget or machine resources.


Sometime, the girlfriend may be busy, need a GPTCache to ease visit stress. :}


Maybe GPTCache can reduce the count of llm request for lower cost. detail: https://github.com/zilliztech/GPTCache


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You