Edit model card

SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2

This is a sentence-transformers model finetuned from sentence-transformers/all-MiniLM-L6-v2. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

Model Details

Model Description

  • Model Type: Sentence Transformer
  • Base model: sentence-transformers/all-MiniLM-L6-v2
  • Maximum Sequence Length: 256 tokens
  • Output Dimensionality: 384 tokens
  • Similarity Function: Cosine Similarity

Model Sources

Full Model Architecture

SentenceTransformer(
  (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel 
  (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
  (2): Normalize()
)

Usage

Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

pip install -U sentence-transformers

Then you can load this model and run inference.

from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("danicafisher/dfisher-sentence-transformer-fine-tuned2")
# Run inference
sentences = [
    'What is the title of the publication related to Artificial Intelligence Risk Management by NIST?',
    'NIST Trustworthy and Responsible AI  \nNIST AI 600-1 \nArtificial Intelligence Risk Management \nFramework: Generative Artificial \nIntelligence Profile \n \n \n \nThis publication is available free of charge from: \nhttps://doi.org/10.6028/NIST.AI.600-1',
    'HUMAN ALTERNATIVES, \nCONSIDERATION, AND \nFALLBACK \nWHAT SHOULD BE EXPECTED OF AUTOMATED SYSTEMS\nThe expectations for automated systems are meant to serve as a blueprint for the development of additional \ntechnical standards and practices that are tailored for particular sectors and contexts. \nAn automated system should provide demonstrably effective mechanisms to opt out in favor of a human alterna\xad\ntive, where appropriate, as well as timely human consideration and remedy by a fallback system, with additional \nhuman oversight and safeguards for systems used in sensitive domains, and with training and assessment for any \nhuman-based portions of the system to ensure effectiveness. \nProvide a mechanism to conveniently opt out from automated systems in favor of a human \nalternative, where appropriate \nBrief, clear, accessible notice and instructions. Those impacted by an automated system should be \ngiven a brief, clear notice that they are entitled to opt-out, along with clear instructions for how to opt-out. \nInstructions should be provided in an accessible form and should be easily findable by those impacted by the \nautomated system. The brevity, clarity, and accessibility of the notice and instructions should be assessed (e.g., \nvia user experience research). \nHuman alternatives provided when appropriate. In many scenarios, there is a reasonable expectation \nof human involvement in attaining rights, opportunities, or access. When automated systems make up part of \nthe attainment process, alternative timely human-driven processes should be provided. The use of a human \nalternative should be triggered by an opt-out process. \nTimely and not burdensome human alternative. Opting out should be timely and not unreasonably \nburdensome in both the process of requesting to opt-out and the human-driven alternative provided. \nProvide timely human consideration and remedy by a fallback and escalation system in the \nevent that an automated system fails, produces error, or you would like to appeal or con\xad\ntest its impacts on you \nProportionate. The availability of human consideration and fallback, along with associated training and \nsafeguards against human bias, should be proportionate to the potential of the automated system to meaning\xad\nfully impact rights, opportunities, or access. Automated systems that have greater control over outcomes, \nprovide input to high-stakes decisions, relate to sensitive domains, or otherwise have a greater potential to \nmeaningfully impact rights, opportunities, or access should have greater availability (e.g., staffing) and over\xad\nsight of human consideration and fallback mechanisms. \nAccessible. Mechanisms for human consideration and fallback, whether in-person, on paper, by phone, or \notherwise provided, should be easy to find and use. These mechanisms should be tested to ensure that users \nwho have trouble with the automated system are able to use human consideration and fallback, with the under\xad\nstanding that it may be these users who are most likely to need the human assistance. Similarly, it should be \ntested to ensure that users with disabilities are able to find and use human consideration and fallback and also \nrequest reasonable accommodations or modifications. \nConvenient. Mechanisms for human consideration and fallback should not be unreasonably burdensome as \ncompared to the automated system’s equivalent. \n49',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]

# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]

Training Details

Training Dataset

Unnamed Dataset

  • Size: 180 training samples
  • Columns: sentence_0 and sentence_1
  • Approximate statistics based on the first 180 samples:
    sentence_0 sentence_1
    type string string
    details
    • min: 12 tokens
    • mean: 22.28 tokens
    • max: 36 tokens
    • min: 21 tokens
    • mean: 241.8 tokens
    • max: 256 tokens
  • Samples:
    sentence_0 sentence_1
    What concerns have been raised regarding the use of facial recognition technology in public housing? 65. See, e.g., Scott Ikeda. Major Data Broker Exposes 235 Million Social Media Profiles in Data Lead: Info
    Appears to Have Been Scraped Without Permission. CPO Magazine. Aug. 28, 2020. https://
    www.cpomagazine.com/cyber-security/major-data-broker-exposes-235-million-social-media-profiles­
    in-data-leak/; Lily Hay Newman. 1.2 Billion Records Found Exposed Online in a Single Server. WIRED,
    Nov. 22, 2019. https://www.wired.com/story/billion-records-exposed-online/
    66. Lola Fadulu. Facial Recognition Technology in Public Housing Prompts Backlash. New York Times.
    Sept. 24, 2019.
    https://www.nytimes.com/2019/09/24/us/politics/facial-recognition-technology-housing.html
    67. Jo Constantz. ‘They Were Spying On Us’: Amazon, Walmart, Use Surveillance Technology to Bust
    Unions. Newsweek. Dec. 13, 2021.
    https://www.newsweek.com/they-were-spying-us-amazon-walmart-use-surveillance-technology-bust­
    unions-1658603
    68. See, e.g., enforcement actions by the FTC against the photo storage app Everalbaum
    (https://www.ftc.gov/legal-library/browse/cases-proceedings/192-3172-everalbum-inc-matter), and
    against Weight Watchers and their subsidiary Kurbo
    (https://www.ftc.gov/legal-library/browse/cases-proceedings/1923228-weight-watchersww)
    69. See, e.g., HIPAA, Pub. L 104-191 (1996); Fair Debt Collection Practices Act (FDCPA), Pub. L. 95-109
    (1977); Family Educational Rights and Privacy Act (FERPA) (20 U.S.C. § 1232g), Children's Online
    Privacy Protection Act of 1998, 15 U.S.C. 6501–6505, and Confidential Information Protection and
    Statistical Efficiency Act (CIPSEA) (116 Stat. 2899)
    70. Marshall Allen. You Snooze, You Lose: Insurers Make The Old Adage Literally True. ProPublica. Nov.
    21, 2018.
    https://www.propublica.org/article/you-snooze-you-lose-insurers-make-the-old-adage-literally-true
    71. Charles Duhigg. How Companies Learn Your Secrets. The New York Times. Feb. 16, 2012.
    https://www.nytimes.com/2012/02/19/magazine/shopping-habits.html
    72. Jack Gillum and Jeff Kao. Aggression Detectors: The Unproven, Invasive Surveillance Technology
    Schools are Using to Monitor Students. ProPublica. Jun. 25, 2019.
    https://features.propublica.org/aggression-detector/the-unproven-invasive-surveillance-technology­
    schools-are-using-to-monitor-students/
    73. Drew Harwell. Cheating-detection companies made millions during the pandemic. Now students are
    fighting back. Washington Post. Nov. 12, 2020.
    https://www.washingtonpost.com/technology/2020/11/12/test-monitoring-student-revolt/
    74. See, e.g., Heather Morrison. Virtual Testing Puts Disabled Students at a Disadvantage. Government
    Technology. May 24, 2022.
    https://www.govtech.com/education/k-12/virtual-testing-puts-disabled-students-at-a-disadvantage;
    Lydia X. Z. Brown, Ridhi Shetty, Matt Scherer, and Andrew Crawford. Ableism And Disability
    Discrimination In New Surveillance Technologies: How new surveillance technologies in education,
    policing, health care, and the workplace disproportionately harm disabled people. Center for Democracy
    and Technology Report. May 24, 2022.
    https://cdt.org/insights/ableism-and-disability-discrimination-in-new-surveillance-technologies-how­
    new-surveillance-technologies-in-education-policing-health-care-and-the-workplace­
    disproportionately-harm-disabled-people/
    69
    What are the potential consequences of automated systems making decisions without providing notice or explanations to affected individuals? NOTICE &
    EXPLANATION
    WHY THIS PRINCIPLE IS IMPORTANT
    This section provides a brief summary of the problems which the principle seeks to address and protect
    against, including illustrative examples.
    Automated systems now determine opportunities, from employment to credit, and directly shape the American
    public’s experiences, from the courtroom to online classrooms, in ways that profoundly impact people’s lives. But this
    expansive impact is not always visible. An applicant might not know whether a person rejected their resume or a
    hiring algorithm moved them to the bottom of the list. A defendant in the courtroom might not know if a judge deny­
    ing their bail is informed by an automated system that labeled them “high risk.” From correcting errors to contesting
    decisions, people are often denied the knowledge they need to address the impact of automated systems on their lives.
    Notice and explanations also serve an important safety and efficacy purpose, allowing experts to verify the reasonable­
    ness of a recommendation before enacting it.
    In order to guard against potential harms, the American public needs to know if an automated system is being used.
    Clear, brief, and understandable notice is a prerequisite for achieving the other protections in this framework. Like­
    wise, the public is often unable to ascertain how or why an automated system has made a decision or contributed to a
    particular outcome. The decision-making processes of automated systems tend to be opaque, complex, and, therefore,
    unaccountable, whether by design or by omission. These factors can make explanations both more challenging and
    more important, and should not be used as a pretext to avoid explaining important decisions to the people impacted
    by those choices. In the context of automated systems, clear and valid explanations should be recognized as a baseline
    requirement.
    Providing notice has long been a standard practice, and in many cases is a legal requirement, when, for example,
    making a video recording of someone (outside of a law enforcement or national security context). In some cases, such
    as credit, lenders are required to provide notice and explanation to consumers. Techniques used to automate the
    process of explaining such systems are under active research and improvement and such explanations can take many
    forms. Innovative companies and researchers are rising to the challenge and creating and deploying explanatory
    systems that can help the public better understand decisions that impact them.
    While notice and explanation requirements are already in place in some sectors or situations, the American public
    deserve to know consistently and across sectors if an automated system is being used in a way that impacts their rights,
    opportunities, or access. This knowledge should provide confidence in how the public is being treated, and trust in the
    validity and reasonable use of automated systems.

    A lawyer representing an older client with disabilities who had been cut off from Medicaid-funded home
    health-care assistance couldn't determine why, especially since the decision went against historical access
    practices. In a court hearing, the lawyer learned from a witness that the state in which the older client
    lived had recently adopted a new algorithm to determine eligibility.83 The lack of a timely explanation made it
    harder to understand and contest the decision.

    A formal child welfare investigation is opened against a parent based on an algorithm and without the parent
    ever being notified that data was being collected and used as part of an algorithmic child maltreatment
    risk assessment.84 The lack of notice or an explanation makes it harder for those performing child
    maltreatment assessments to validate the risk assessment and denies parents knowledge that could help them
    contest a decision.
    41
    How has the Supreme Court's decision to overturn Roe v Wade been addressed by President Biden? ENDNOTES
    1.The Executive Order On Advancing Racial Equity and Support for Underserved Communities Through the
    Federal Government. https://www.whitehouse.gov/briefing-room/presidential-actions/2021/01/20/executive
    order-advancing-racial-equity-and-support-for-underserved-communities-through-the-federal-government/
    2. The White House. Remarks by President Biden on the Supreme Court Decision to Overturn Roe v. Wade. Jun.
    24, 2022. https://www.whitehouse.gov/briefing-room/speeches-remarks/2022/06/24/remarks-by-president­
    biden-on-the-supreme-court-decision-to-overturn-roe-v-wade/
    3. The White House. Join the Effort to Create A Bill of Rights for an Automated Society. Nov. 10, 2021. https://
    www.whitehouse.gov/ostp/news-updates/2021/11/10/join-the-effort-to-create-a-bill-of-rights-for-an­
    automated-society/
    4. U.S. Dept. of Health, Educ. & Welfare, Report of the Sec’y’s Advisory Comm. on Automated Pers. Data Sys.,
    Records, Computers, and the Rights of Citizens (July 1973). https://www.justice.gov/opcl/docs/rec-com­
    rights.pdf.
    5. See, e.g., Office of Mgmt. & Budget, Exec. Office of the President, Circular A-130, Managing Information as a
    Strategic Resource, app. II § 3 (July 28, 2016); Org. of Econ. Co-Operation & Dev., Revision of the
    Recommendation of the Council Concerning Guidelines Governing the Protection of Privacy and Transborder
    Flows of Personal Data, Annex Part Two (June 20, 2013). https://one.oecd.org/document/C(2013)79/en/pdf.
    6. Andrew Wong et al. External validation of a widely implemented proprietary sepsis prediction model in
    hospitalized patients. JAMA Intern Med. 2021; 181(8):1065-1070. doi:10.1001/jamainternmed.2021.2626
    7. Jessica Guynn. Facebook while black: Users call it getting 'Zucked,' say talking about racism is censored as hate
    speech. USA Today. Apr. 24, 2019. https://www.usatoday.com/story/news/2019/04/24/facebook-while-black­
    zucked-users-say-they-get-blocked-racism-discussion/2859593002/
    8. See, e.g., Michael Levitt. AirTags are being used to track people and cars. Here's what is being done about it.
    NPR. Feb. 18, 2022. https://www.npr.org/2022/02/18/1080944193/apple-airtags-theft-stalking-privacy-tech;
    Samantha Cole. Police Records Show Women Are Being Stalked With Apple AirTags Across the Country.
    Motherboard. Apr. 6, 2022. https://www.vice.com/en/article/y3vj3y/apple-airtags-police-reports-stalking­
    harassment
    9. Kristian Lum and William Isaac. To Predict and Serve? Significance. Vol. 13, No. 5, p. 14-19. Oct. 7, 2016.
    https://rss.onlinelibrary.wiley.com/doi/full/10.1111/j.1740-9713.2016.00960.x; Aaron Sankin, Dhruv Mehrotra,
    Surya Mattu, and Annie Gilbertson. Crime Prediction Software Promised to Be Free of Biases. New Data Shows
    It Perpetuates Them. The Markup and Gizmodo. Dec. 2, 2021. https://themarkup.org/prediction­
    bias/2021/12/02/crime-prediction-software-promised-to-be-free-of-biases-new-data-shows-it-perpetuates­
    them
    10. Samantha Cole. This Horrifying App Undresses a Photo of Any Woman With a Single Click. Motherboard.
    June 26, 2019. https://www.vice.com/en/article/kzm59x/deepnude-app-creates-fake-nudes-of-any-woman
    11. Lauren Kaori Gurley. Amazon’s AI Cameras Are Punishing Drivers for Mistakes They Didn’t Make.
    Motherboard. Sep. 20, 2021. https://www.vice.com/en/article/88npjv/amazons-ai-cameras-are-punishing­
    drivers-for-mistakes-they-didnt-make
    63
  • Loss: MultipleNegativesRankingLoss with these parameters:
    {
        "scale": 20.0,
        "similarity_fct": "cos_sim"
    }
    

Training Hyperparameters

Non-Default Hyperparameters

  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • multi_dataset_batch_sampler: round_robin

All Hyperparameters

Click to expand
  • overwrite_output_dir: False
  • do_predict: False
  • eval_strategy: no
  • prediction_loss_only: True
  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • per_gpu_train_batch_size: None
  • per_gpu_eval_batch_size: None
  • gradient_accumulation_steps: 1
  • eval_accumulation_steps: None
  • torch_empty_cache_steps: None
  • learning_rate: 5e-05
  • weight_decay: 0.0
  • adam_beta1: 0.9
  • adam_beta2: 0.999
  • adam_epsilon: 1e-08
  • max_grad_norm: 1
  • num_train_epochs: 3
  • max_steps: -1
  • lr_scheduler_type: linear
  • lr_scheduler_kwargs: {}
  • warmup_ratio: 0.0
  • warmup_steps: 0
  • log_level: passive
  • log_level_replica: warning
  • log_on_each_node: True
  • logging_nan_inf_filter: True
  • save_safetensors: True
  • save_on_each_node: False
  • save_only_model: False
  • restore_callback_states_from_checkpoint: False
  • no_cuda: False
  • use_cpu: False
  • use_mps_device: False
  • seed: 42
  • data_seed: None
  • jit_mode_eval: False
  • use_ipex: False
  • bf16: False
  • fp16: False
  • fp16_opt_level: O1
  • half_precision_backend: auto
  • bf16_full_eval: False
  • fp16_full_eval: False
  • tf32: None
  • local_rank: 0
  • ddp_backend: None
  • tpu_num_cores: None
  • tpu_metrics_debug: False
  • debug: []
  • dataloader_drop_last: False
  • dataloader_num_workers: 0
  • dataloader_prefetch_factor: None
  • past_index: -1
  • disable_tqdm: False
  • remove_unused_columns: True
  • label_names: None
  • load_best_model_at_end: False
  • ignore_data_skip: False
  • fsdp: []
  • fsdp_min_num_params: 0
  • fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
  • fsdp_transformer_layer_cls_to_wrap: None
  • accelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
  • deepspeed: None
  • label_smoothing_factor: 0.0
  • optim: adamw_torch
  • optim_args: None
  • adafactor: False
  • group_by_length: False
  • length_column_name: length
  • ddp_find_unused_parameters: None
  • ddp_bucket_cap_mb: None
  • ddp_broadcast_buffers: False
  • dataloader_pin_memory: True
  • dataloader_persistent_workers: False
  • skip_memory_metrics: True
  • use_legacy_prediction_loop: False
  • push_to_hub: False
  • resume_from_checkpoint: None
  • hub_model_id: None
  • hub_strategy: every_save
  • hub_private_repo: False
  • hub_always_push: False
  • gradient_checkpointing: False
  • gradient_checkpointing_kwargs: None
  • include_inputs_for_metrics: False
  • eval_do_concat_batches: True
  • fp16_backend: auto
  • push_to_hub_model_id: None
  • push_to_hub_organization: None
  • mp_parameters:
  • auto_find_batch_size: False
  • full_determinism: False
  • torchdynamo: None
  • ray_scope: last
  • ddp_timeout: 1800
  • torch_compile: False
  • torch_compile_backend: None
  • torch_compile_mode: None
  • dispatch_batches: None
  • split_batches: None
  • include_tokens_per_second: False
  • include_num_input_tokens_seen: False
  • neftune_noise_alpha: None
  • optim_target_modules: None
  • batch_eval_metrics: False
  • eval_on_start: False
  • eval_use_gather_object: False
  • batch_sampler: batch_sampler
  • multi_dataset_batch_sampler: round_robin

Framework Versions

  • Python: 3.11.9
  • Sentence Transformers: 3.1.1
  • Transformers: 4.44.2
  • PyTorch: 2.4.1
  • Accelerate: 0.34.2
  • Datasets: 3.0.0
  • Tokenizers: 0.19.1

Citation

BibTeX

Sentence Transformers

@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}

MultipleNegativesRankingLoss

@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
Downloads last month
9
Safetensors
Model size
22.7M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for danicafisher/dfisher-sentence-transformer-fine-tuned2

Finetuned
(162)
this model

Space using danicafisher/dfisher-sentence-transformer-fine-tuned2 1