| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
| |
Now the API itself is responsible for querying settings. This makes
sense, as it's an internal part of the component.
|
| |
|
|
|
|
|
|
|
|
|
| |
There's no preferences dialog, so you can't really adjust the prompt
or the model it uses. The default settings work well for me. You may
want to tweak them depending on your model preferences and compute
budget. (Not many can afford to run Llama3-8B at high
quantization. Conversely, you might have a better GPU than me and wish
to run a 27B model or bigger.)
|
|
|
|
|
|
|
|
|
|
|
| |
1. narrow_layout is indeed unused, `AdwBreakpoint` remembers initial
values and restores them on load
2. `TagPillWidgets` is not exactly dead code, I'm adding all the
widgets to it to emulate how Relm4 macros do it. Perhaps it'll be used
in the future.
3. We currently ignore the libsecret result on clearing the tokens
because dealing with libsecret errors is annoying and requires unsafe
code due to outdated dependencies.
|
|
|
|
|
|
| |
This is a very shitty translation, but it can be improved later. I
added it mostly as a test for translations working correctly, since I
know Russian and might as well translate the app into the language.
|
| |
|
| |
|
|
|
|
| |
Yes, I'm stupid and I'm developing under a custom theme
|
| |
|
| |
|
|
|
|
| |
TODO: make requested scopes parametrized
|
|
|
|
|
|
|
| |
This helps unify HTTP-related settings in one place.
In Relm4 Matrix chatroom, it was said that deconstructing child
components makes little sense in terms of optimizations.
|
|
|
|
|
|
| |
The code is really janky and unpolished, the error handling is
TERRIBLE, and I think I can't publish it like this. This'll need a
refactor, but it'll come tomorrow.
|
| |
|
| |
|
|
|
|
|
| |
Now it's easy to use the same UI for sending a new post or editing an
existing one (by loading it with `?q=source` and then comparing).
|
|
|
|
|
|
|
| |
What this command should do is construct a summarization request and
return a future which would return chunks from the LLM.
Perhaps this component will be asyncified in the future.
|
|
|
|
|
| |
On receving `smart_summary::Output::Start`, one must reply with
`smart_summary::Input::Text(text)` to start the actual summarization.
|
|
This is a little bit janky in my opinion, because it takes a reference
to the buffer which contents its gonna be summarizing. In a perfect
world, it would ask the parent component for the text.
|