feat: Ollama keep_alive setting, stream stop, and API hardening
- Options: configurable model keep-alive (Ollama native) with labeled presets
- Popup: chat send becomes Stop during replies; abort covers pre-stream wait
- Background: per-tab stream/other abort slots, batched stream chunks, Connection: close
- Config: stream caps, suggestion token limits; pass keepAlive through all Ollama calls