Commit
Β·
2dd100c
1
Parent(s):
c293902
docs: clarify title handling
Browse files
README.md
CHANGED
|
@@ -100,7 +100,7 @@ Key parameters you may want to tune:
|
|
| 100 |
|
| 101 |
- **`question: str | Sequence[str]`** β Query text. Provide a list to batch multiple questions; each item pairs with the corresponding entry in `context`.
|
| 102 |
- **`context: str | Sequence[str] | Sequence[Sequence[str]]`** β Contexts aligned to the query. Use a list for one document per query, or a list of lists to supply multiple documents (or pre-split sentences) for each query.
|
| 103 |
-
- **`title: str | Sequence[str] | Sequence[Sequence[str]] | None`** β Optional titles aligned to each context. The default `"first_sentence"`
|
| 104 |
- **`threshold: float` (default `0.1`)** β Pruning probability cutoff. Larger values discard more sentences; `0.05β0.5` works well across datasets.
|
| 105 |
- **`batch_size: int` (default `32`)** β Number of contexts processed per inference batch. Increase for throughput, decrease if you run out of memory.
|
| 106 |
- **`language: str | None`** β Built-in splitter selection (`"auto"`, `"ja"`, `"en"`). The default behaves like `"auto"` and detects Japanese vs. English automatically.
|
|
|
|
| 100 |
|
| 101 |
- **`question: str | Sequence[str]`** β Query text. Provide a list to batch multiple questions; each item pairs with the corresponding entry in `context`.
|
| 102 |
- **`context: str | Sequence[str] | Sequence[Sequence[str]]`** β Contexts aligned to the query. Use a list for one document per query, or a list of lists to supply multiple documents (or pre-split sentences) for each query.
|
| 103 |
+
- **`title: str | Sequence[str] | Sequence[Sequence[str]] | None`** β Optional titles aligned to each context. The default sentinel `"first_sentence"` marks the opening sentence so you can keep it by pairing with `always_select_title=True` or `first_line_as_title=True`; without those flags it is scored like any other sentence. Set `None` to disable all title handling.
|
| 104 |
- **`threshold: float` (default `0.1`)** β Pruning probability cutoff. Larger values discard more sentences; `0.05β0.5` works well across datasets.
|
| 105 |
- **`batch_size: int` (default `32`)** β Number of contexts processed per inference batch. Increase for throughput, decrease if you run out of memory.
|
| 106 |
- **`language: str | None`** β Built-in splitter selection (`"auto"`, `"ja"`, `"en"`). The default behaves like `"auto"` and detects Japanese vs. English automatically.
|