Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
Paper • 1908.10084 • Published • 13
How to use yasserrmd/pediatrics-gemma-300m-emb with sentence-transformers:
from sentence_transformers import SentenceTransformer
model = SentenceTransformer("yasserrmd/pediatrics-gemma-300m-emb")
sentences = [
"What are the main clinical manifestations and characteristics of congenital syphilis?",
"Trained staff in community health centers obtain information on infant weight during periodic visits scheduled at three and six months. Age- and sex-adjusted standard deviation scores (SDS) for weight are calculated using Dutch reference curves. Infant growth rates of weight gain in the first three and six months after birth are then calculated by comparing the weight at three or six months to the birth weight SDS. These growth rates provide information on the infant's growth trajectory and can help identify any deviations from the norm.",
"The main clinical manifestations of congenital syphilis include prematurity, low birth weight, hepatomegaly with or without splenomegaly, cutaneous lesions, limb pseudoparalysis, respiratory distress, jaundice, anemia, generalized lymphadenopathy, and osteitis and osteochondritis. Late congenital syphilis is characterized by saber shin deformity of the tibia, Clutton's joints, frontal bossing, saddle nose, deformed upper medial incisor teeth (Hutchinson's teeth), neurological deafness, and difficulty in learning.",
"When evaluating newborns, orthopaedic surgeons should consider risk factors such as difficult vaginal delivery of a large infant, which can be associated with fractures or brachial plexus injury, and breech position, which can be associated with developmental dysplasia of the hip (DDH). They should also be aware of conditions like Klippel-Feil syndrome, congenital muscular torticollis, arthrogryposis, and neurologic problems indicated by a high arch in the feet."
]
embeddings = model.encode(sentences)
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [4, 4]This is a sentence-transformers model finetuned from google/embeddinggemma-300m. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
SentenceTransformer(
(0): Transformer({'max_seq_length': 2048, 'do_lower_case': False, 'architecture': 'Gemma3TextModel'})
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Dense({'in_features': 768, 'out_features': 3072, 'bias': False, 'activation_function': 'torch.nn.modules.linear.Identity'})
(3): Dense({'in_features': 3072, 'out_features': 768, 'bias': False, 'activation_function': 'torch.nn.modules.linear.Identity'})
(4): Normalize()
)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("yasserrmd/pediatrics-gemma-300m-emb")
# Run inference
queries = [
"How can interdisciplinary collaboration and clearly defined strategies help reduce damage and suffering for children and adolescents with mental disorders and their families?\n",
]
documents = [
'Interdisciplinary collaboration, involving caregivers, mental health managers, and professionals, can play a crucial role in reducing damage and suffering for children and adolescents with mental disorders and their families. By developing a shared vision and practice, professionals can work together to establish dynamic strategies that address the complex demands of these individuals. This collaboration requires critical observation of services, conduct, and strategies to ensure that families and mental health services are brought closer together. By bridging the gap between demands and care, interdisciplinary collaboration can help improve the well-being and outcomes for children, adolescents, and their families.',
'Pneumothorax in children can have various causes, including lung cysts. In cases where there are signs of pneumothorax without a lung lesion to account for the condition, cysts of the lung should be suspected. The presence of cysts can be confirmed through radiographic evidence, which can also help determine the location and characteristics of the cysts.',
'The infant in the NICU presented with a degraded general status, intubation, and mechanical ventilation. They also had unilaterally diminished breath sounds, hypoxemia, oliguria, tachycardia, hypotension, abdominal distention, and fever. Additionally, they exhibited hepatosplenomegaly, oliguria progressing to anuria, thrombocytopenia, and hyperchromic hematuria. ',
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 768] [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[ 0.7673, -0.0352, 0.0221]])
sentence_0 and sentence_1| sentence_0 | sentence_1 | |
|---|---|---|
| type | string | string |
| details |
|
|
| sentence_0 | sentence_1 |
|---|---|
What is the role of routine health check-ups in detecting and diagnosing metabolic syndrome and NAFLD in obese children? |
Routine health check-ups are important in detecting and diagnosing metabolic syndrome and NAFLD in obese children. However, there is a lack of routine health check-up data specifically for these complications in obese children. To address this need, pediatric health promotion centers and pediatric obesity clinics have been developed. The aim of these centers is to provide routine health check-ups and obesity-oriented check-ups to detect and diagnose metabolic syndrome and NAFLD in children. |
How does the implementation of family-centered rounds (FCR) impact medical education? |
The implementation of family-centered rounds (FCR) has raised concerns about its potential impact on medical education. Some evidence suggests that FCR may lead to decreased "didactic" teaching, increased discomfort in asking specific management questions, and limited time to discuss management options for residents and students (8, 9, 10, 11). However, the literature on the association between FCR and teaching has mainly focused on learners' perceptions, and there is a lack of objective data to address the relationship between FCR and medical knowledge acquisition. |
What are some common clinical symptoms of neonatal septicaemia? |
Some common clinical symptoms of neonatal septicaemia include fever, poor feeding, excessive cry, difficulty in breathing, yellowish skin discoloration, skin rashes, jitteriness, and irritability. |
MultipleNegativesRankingLoss with these parameters:{
"scale": 20.0,
"similarity_fct": "cos_sim",
"gather_across_devices": false
}
per_device_train_batch_size: 6per_device_eval_batch_size: 6num_train_epochs: 1multi_dataset_batch_sampler: round_robinoverwrite_output_dir: Falsedo_predict: Falseeval_strategy: noprediction_loss_only: Trueper_device_train_batch_size: 6per_device_eval_batch_size: 6per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 5e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1num_train_epochs: 1max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.0warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Falsefp16: Falsefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}parallelism_config: Nonedeepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torch_fusedoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsehub_revision: Nonegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters: auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseliger_kernel_config: Noneeval_use_gather_object: Falseaverage_tokens_across_devices: Falseprompts: Nonebatch_sampler: batch_samplermulti_dataset_batch_sampler: round_robinrouter_mapping: {}learning_rate_mapping: {}| Epoch | Step | Training Loss |
|---|---|---|
| 0.1500 | 500 | 0.0195 |
| 0.2999 | 1000 | 0.0095 |
| 0.4499 | 1500 | 0.0084 |
| 0.5999 | 2000 | 0.0059 |
| 0.7499 | 2500 | 0.0021 |
| 0.8998 | 3000 | 0.0035 |
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Base model
google/embeddinggemma-300m