matthewdicks98 commited on
Commit
c8a7a96
·
verified ·
1 Parent(s): 21a893d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +137 -129
README.md CHANGED
@@ -1,129 +1,137 @@
1
- ---
2
- license: odc-by
3
- task_categories:
4
- - text-classification
5
- language:
6
- - en
7
- tags:
8
- - tense-prediction
9
- - finance
10
- - nlp
11
- size_categories:
12
- - 100K<n<1M
13
- ---
14
-
15
- <p align="center">
16
- <img src="https://github.com/NosibleAI/nosible-py/blob/main/docs/_static/readme.png?raw=true"/>
17
- <p>
18
-
19
- # NOSIBLE Predictive Dataset
20
-
21
- ## Changelog
22
- - **v1.0.0:** Initial version
23
-
24
- ## Who is NOSIBLE?
25
-
26
- NOSIBLE is a web-scale vertical search engine. Our worldwide media surveillance products help companies build AI systems that see every worldwide event and act with complete situational awareness. In short, we help companies know everything, all the time. The financial institutions we work with rely on us to deliver media intelligence from every country in every language in real-time. Shortcomings in existing financial datasets and financial models are what inspired us to release this dataset and related models.
27
-
28
- - [NOSIBLE Financial Sentiment v1.1 Base](https://huggingface.co/NOSIBLE/financial-sentiment-v1.1-base)
29
- - [NOSIBLE Forward Looking v1.1 Base](https://huggingface.co/NOSIBLE/forward-looking-v1.1-base)
30
- - **[NOSIBLE Prediction v1.1 Base](https://huggingface.co/NOSIBLE/prediction-v1.1-base)**
31
-
32
- ## What is it?
33
-
34
- The NOSIBLE Forward Looking Dataset is an open collection of **100,000** cleaned, deduplicated, and labeled financial news text samples. The data was extracted directly from the NOSIBLE search engine in response to real queries asked by real financial institutions. Each sample has been labeled as **predictive** or **not-predictive**, using a sophisticated labeling pipeline (more information coming soon, and a brief explanation below).
35
-
36
- ## How to use it
37
- Using the [HuggingFace datasets library](https://huggingface.co/docs/datasets/):
38
-
39
- You can install it with `pip install datasets`, and must login using e.g. `hf auth login` to access this dataset.
40
-
41
- ```python
42
- from datasets import load_dataset
43
-
44
- dataset = load_dataset("NOSIBLE/predictive")
45
-
46
- print(dataset)
47
-
48
- # DatasetDict({
49
- # train: Dataset({
50
- # features: ['text', 'label', 'netloc', 'url'],
51
- # num_rows: 100000
52
- # })
53
- # })
54
-
55
- # What's next?
56
- # Train your model 🤖
57
- # Profit 💰
58
- ```
59
-
60
- ## Dataset Structure
61
-
62
- ### Data Instances
63
-
64
- The following is an example sample from the dataset:
65
-
66
- ```json
67
- {
68
- "text": "Offshore staff HOUSTON \u2013 VAALCO Energy is looking to bring in a floating storage and offloading (FSO) unit at the Etame Marin oil field offshore Gabon. The company has signed a non-binding letter of intent with Omni Offshore Terminals to supply and operate the vessel at Etame for up to 11 years, following the expiry of the current FPSO Petr\u00f3leo Nautipa contract with BW Offshore in September 2022. Omni has provided a preliminary proposal for leasing and operating the FSO, which could reduce VAALCO's operating costs by 15-25%, compared with the current FPSO contract during the term of the proposed agreement. Maintaining the current FPSO beyond its contract or transitioning to a different FPSO, VAALCO added, would require substantial capex investments.",
69
- "label": "predictive",
70
- "netloc": "www.offshore-mag.com",
71
- "url": "https://www.offshore-mag.com/rigs-vessels/article/14202267/vaalco-contemplating-switch-to-fso-at-etame-offshore-gabon"
72
- }
73
- ```
74
-
75
- ### Data Fields
76
-
77
- - `text` (string): A text chunk from a document.
78
- - `label` (string): The causal structure label of the text, sourced from LLMs and refined with active learning (an iterative relabeling process).
79
- - `netloc` (string): The network location (domain) of the document.
80
- - `url` (string): The URL of the document.
81
-
82
- ## Dataset creation
83
-
84
- ### Data source
85
- The dataset was sampled from NOSIBLE datafeeds, which provides web-scale surveillance data to customers. Samples consist of top-ranked search results from the NOSIBLE search engine in response to safe, curated queries. All data is sourced exclusively from the public web.
86
-
87
- ### Relabeling algorithm
88
- The dataset's label field was annotated by LLMs and refined using an active learning algorithm called relabeling.
89
-
90
- The algorithm outline is as follows:
91
-
92
- 1. Hand label a candidate set of ~200 samples to use as a test bed to refine the prompt used by the LLM labelers to classify the text.
93
- 2. Label a set of 100k samples with LLM labelers:
94
- - `x-ai/grok-4-fast`
95
- - `x-ai/grok-4-fast:thinking`
96
- - `google/gemini-2.5-flash`
97
- - `openai/gpt-5-nano`
98
- - `openai/gpt-4.1-mini`
99
- - `openai/gpt-oss-120b`
100
- - `meta-llama/llama-4-maverick`
101
- - `qwen/qwen3-32b`
102
- 3. Train a linear model on the labels using the majority vote of the LLM labelers.
103
- 4. Iterative relabeling (active learning steps) to improve the label quality:
104
- - Evaluate linear model's predictions over the samples.
105
- - Find disagreements: samples where the LLM labelers agree on a label, but the model has predicted a different label.
106
- - Consult a much larger LLM, the oracle, to evaluate the model's prediction and relabel the sample if it agrees with the LLM labels.
107
- - Drop the worst performing LLM labelers from the ensemble.
108
- - Repeat the process with the remaining LLM labelers until the number of samples relabeled reaches 0.
109
- - Store the refined relabeled dataset.
110
- 5. This is the final dataset used for training the [NOSIBLE Prediction v1.1 Base](https://huggingface.co/NOSIBLE/prediction-v1.1-base) model, which is a finetune.
111
-
112
- ## Additional information
113
-
114
- ### License
115
- The dataset is released under the **Open Data Commons Attribution License (ODC-By) v1.0** [license](https://opendatacommons.org/licenses/by/1-0/).
116
-
117
- ### Attribution
118
- - [NOSIBLE Inc](https://www.nosible.com/) Team
119
-
120
- **Contributors**
121
-
122
- This dataset was developed and maintained by the following team:
123
-
124
- * [**Matthew Dicks**](https://www.linkedin.com/in/matthewdicks98/)
125
- * [**Simon van Dyk**](https://www.linkedin.com/in/simon-van-dyk/)
126
- * [**Stuart Reid**](https://www.linkedin.com/in/stuartgordonreid/)
127
-
128
- ## Citations
129
- Coming soon, we're working on a white paper.
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: odc-by
3
+ task_categories:
4
+ - text-classification
5
+ language:
6
+ - en
7
+ tags:
8
+ - tense-prediction
9
+ - finance
10
+ - nlp
11
+ size_categories:
12
+ - 100K<n<1M
13
+ ---
14
+
15
+ <p align="center">
16
+ <img src="https://github.com/NosibleAI/nosible-py/blob/main/docs/_static/readme.png?raw=true"/>
17
+ <p>
18
+
19
+ # NOSIBLE Prediction Dataset
20
+
21
+ ## Changelog
22
+ - **v1.0.0:** Initial version
23
+
24
+ ## Who is NOSIBLE?
25
+
26
+ [**NOSIBLE**](https://www.nosible.com/) is a vertical web-scale search engine. Our worldwide media surveillance products help companies build AI systems that see every worldwide event and act with complete situational awareness. In short, we help companies know everything, all the time. The financial institutions we work with rely on us to deliver media intelligence from every country in every language in real-time. Shortcomings in existing financial datasets and financial models are what inspired us to release this dataset and related models.
27
+
28
+ - [**NOSIBLE Prediction v1.1 Base**](https://huggingface.co/NOSIBLE/prediction-v1.1-base)
29
+
30
+ ## What is it?
31
+
32
+ The NOSIBLE Prediction Dataset is an open collection of **100,000** cleaned and deduplicated news samples curated to identify predictive language. Each sample is annotated with one of two labels: **prediction**, indicating that the text includes a prediction, estimate, or forecast, and **not-prediction**, indicating that it does not.
33
+
34
+ All text is sourced from the **NOSIBLE Search Feeds** product using a curated set of finance-related queries. Sentiment labels are assigned through a multi-stage, LLM-based annotation pipeline (described below).
35
+
36
+ ## How to use it
37
+ Using the [HuggingFace datasets library](https://huggingface.co/docs/datasets/):
38
+
39
+ Install the dataset library with `pip install datasets`, then load the dataset:
40
+
41
+ ```python
42
+ from datasets import load_dataset
43
+
44
+ dataset = load_dataset("NOSIBLE/prediction")
45
+ print(dataset)
46
+ ```
47
+
48
+ #### Expected Output
49
+
50
+ ```text
51
+ DatasetDict({
52
+ train: Dataset({
53
+ features: ['text', 'label', 'netloc', 'url'],
54
+ num_rows: 100000
55
+ })
56
+ })
57
+ ```
58
+
59
+ You can also access this dataset through any interface supported by [Hugging Face](https://huggingface.co/).
60
+
61
+ ## Dataset Structure
62
+
63
+ ### Data Instances
64
+
65
+ The following is an example sample from the dataset:
66
+
67
+ ```json
68
+ {
69
+ "text": "Offshore staff HOUSTON \u2013 VAALCO Energy is looking to bring in a floating storage and offloading (FSO) unit at the Etame Marin oil field offshore Gabon. The company has signed a non-binding letter of intent with Omni Offshore Terminals to supply and operate the vessel at Etame for up to 11 years, following the expiry of the current FPSO Petr\u00f3leo Nautipa contract with BW Offshore in September 2022. Omni has provided a preliminary proposal for leasing and operating the FSO, which could reduce VAALCO's operating costs by 15-25%, compared with the current FPSO contract during the term of the proposed agreement. Maintaining the current FPSO beyond its contract or transitioning to a different FPSO, VAALCO added, would require substantial capex investments.",
70
+ "label": "prediction",
71
+ "netloc": "www.offshore-mag.com",
72
+ "url": "https://www.offshore-mag.com/rigs-vessels/article/14202267/vaalco-contemplating-switch-to-fso-at-etame-offshore-gabon"
73
+ }
74
+ ```
75
+
76
+ ### Data Fields
77
+
78
+ - `text` (string): A text chunk from a search result.
79
+ - `label` (string): The causal structure label.
80
+ - `netloc` (string): The domain name of the source document.
81
+ - `url` (string): The URL of the document.
82
+
83
+ ## Dataset creation
84
+
85
+ ### Data source
86
+ The dataset was sampled from the NOSIBLE Search Feeds, which provides web-scale surveillance data to customers. Samples consist of top-ranked search results from the NOSIBLE search engine in response to safe, curated, and finance-specific queries. All data is sourced exclusively from the public web.
87
+
88
+ ### Relabeling algorithm
89
+ Labels were first generated using multiple LLM annotators and were then refined using an active-learning–based relabeling loop.
90
+
91
+ The algorithm outline is as follows:
92
+
93
+ 1. Hand-label ~200 samples to tune prompts for the LLM annotators.
94
+ 2. Label a set of 100k samples with LLM labelers:
95
+ - [`xAI: Grok 4 Fast`](https://openrouter.ai/x-ai/grok-4-fast)
96
+ - [`xAI: Grok 4 Fast (reasoning enabled)`](https://openrouter.ai/x-ai/grok-4-fast)
97
+ - [`Google: Gemini 2.5 Flash`](https://openrouter.ai/google/gemini-2.5-flash)
98
+ - [`OpenAI: GPT-5 Nano`](https://openrouter.ai/openai/gpt-5-nano)
99
+ - [`OpenAI: GPT-4.1 Mini`](https://openrouter.ai/openai/gpt-4.1-mini)
100
+ - [`OpenAI: gpt-oss-120b`](https://openrouter.ai/openai/gpt-oss-120b)
101
+ - [`Meta: Llama 4 Maverick`](https://openrouter.ai/meta-llama/llama-4-maverick)
102
+ - [`Qwen: Qwen3 32B`](https://openrouter.ai/qwen/qwen3-32b)
103
+ 3. Train multiple linear models to predict the majority-vote of the LLM labelers. The features are the text embeddings of the following models:
104
+ - [`Qwen3-Embedding-8B`](https://openrouter.ai/qwen/qwen3-embedding-8b)
105
+ - [`Qwen3-Embedding-4B`](https://openrouter.ai/qwen/qwen3-embedding-4b)
106
+ - [`Qwen3-Embedding-0.6B`](https://openrouter.ai/qwen/qwen3-embedding-0.6b)
107
+ - [`OpenAI: Text Embedding 3 Large`](https://openrouter.ai/openai/text-embedding-3-large)
108
+ - [`Google: Gemini Embedding 001`](https://openrouter.ai/google/gemini-embedding-001)
109
+ - [`Mistral: Mistral Embed 2312`](https://openrouter.ai/mistralai/mistral-embed-2312)
110
+ 4. Perform iterative relabeling:
111
+ - Compare all the linear models' predictions to the majority-vote label.
112
+ - Identify disagreements where all linear models agree but the majority-vote label does not.
113
+ - Use a larger LLM (“oracle”) to evaluate ambiguous cases and relabel when appropriate.
114
+ - Drop the worst performing linear models from the ensemble.
115
+ - Repeat until no additional samples require relabeling.
116
+ 5. This is the final dataset used for training the [NOSIBLE Prediction v1.1 Base](https://huggingface.co/NOSIBLE/prediction-v1.1-base) model.
117
+
118
+ We used [`OpenAI: GPT-5.1`](https://openrouter.ai/openai/gpt-5.1) as the oracle.
119
+
120
+ ## Additional information
121
+
122
+ ### License
123
+ The dataset is released under the **Open Data Commons Attribution License (ODC-By) v1.0** [license](https://opendatacommons.org/licenses/by/1-0/).
124
+
125
+ ### Attribution
126
+ - [NOSIBLE Inc](https://www.nosible.com/) Team
127
+
128
+ **Contributors**
129
+
130
+ This dataset was developed and maintained by the following team:
131
+
132
+ * [**Matthew Dicks**](https://www.linkedin.com/in/matthewdicks98/)
133
+ * [**Simon van Dyk**](https://www.linkedin.com/in/simon-van-dyk/)
134
+ * [**Stuart Reid**](https://www.linkedin.com/in/stuartgordonreid/)
135
+
136
+ ## Citations
137
+ Coming soon, we're working on a white paper.