Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some my question about GPT4all tool configuration settings #357

Open
jasonsu123 opened this issue Nov 27, 2024 · 0 comments
Open

Some my question about GPT4all tool configuration settings #357

jasonsu123 opened this issue Nov 27, 2024 · 0 comments

Comments

@jasonsu123
Copy link

Hello everyone,

I’m currently using the latest version of GPT4all, and I must say it’s one of the easiest and most intuitive tools for running local LLMs. However, I’ve encountered some questions during my use:

GPT4all offers an API server feature. Does this mean I can use my OpenAI API key to connect to OpenAI’s server models? I noticed the API server is on port 4891, but when I try to access http://localhost:4891/ via a browser, it shows that the connection cannot be established.

In the settings under model configuration, I see options for context length and max token length. Are these adjustable, or are they fixed at 2048 and 4096? Or do these values change based on the GGUF model being loaded?

Additionally, I’m unclear about what context length means. Does it indicate that the model will only consider 2048 tokens at a time? And is the maximum always capped at 4096?

If I’m using a model designed for handling long documents, such as Qwen-2.5, how should I configure the context length and max token length settings?

Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant