modelId
stringlengths
5
122
author
stringlengths
2
42
last_modified
unknown
downloads
int64
0
167M
likes
int64
0
6.49k
library_name
stringclasses
331 values
tags
sequencelengths
1
4.05k
pipeline_tag
stringclasses
51 values
createdAt
unknown
card
stringlengths
1
913k
leogomezr/leonelgomez
leogomezr
"2024-09-25T00:09:10Z"
0
0
null
[ "license:other", "region:us" ]
null
"2024-09-24T23:07:30Z"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
xueyj/Qwen-Qwen1.5-7B-1727219303
xueyj
"2024-09-24T23:08:34Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-7B", "base_model:adapter:Qwen/Qwen1.5-7B", "region:us" ]
null
"2024-09-24T23:08:23Z"
--- base_model: Qwen/Qwen1.5-7B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
crystaltine/a2c-PandaReachDense-v3
crystaltine
"2024-09-24T23:12:45Z"
0
0
stable-baselines3
[ "stable-baselines3", "PandaReachDense-v3", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
"2024-09-24T23:08:26Z"
--- library_name: stable-baselines3 tags: - PandaReachDense-v3 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: PandaReachDense-v3 type: PandaReachDense-v3 metrics: - type: mean_reward value: -0.23 +/- 0.10 name: mean_reward verified: false --- # **A2C** Agent playing **PandaReachDense-v3** This is a trained model of a **A2C** agent playing **PandaReachDense-v3** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
wdfgrfxthb/LIPITOR-Can-lipitor-affect-muscle-strength-during-exercise-DrugChatter-g1-updated
wdfgrfxthb
"2024-09-24T23:08:38Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:08:36Z"
Entry not found
Krabat/google-gemma-2b-1727219332
Krabat
"2024-09-24T23:08:56Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:08:53Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
jet-taekyo/snowflake_finetuned_recursive
jet-taekyo
"2024-09-24T23:09:27Z"
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:714", "loss:MatryoshkaLoss", "loss:MultipleNegativesRankingLoss", "arxiv:1908.10084", "arxiv:2205.13147", "arxiv:1705.00652", "base_model:Snowflake/snowflake-arctic-embed-m", "base_model:finetune:Snowflake/snowflake-arctic-embed-m", "model-index", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
"2024-09-24T23:08:54Z"
--- base_model: Snowflake/snowflake-arctic-embed-m library_name: sentence-transformers metrics: - cosine_accuracy@1 - cosine_accuracy@3 - cosine_accuracy@5 - cosine_accuracy@10 - cosine_precision@1 - cosine_precision@3 - cosine_precision@5 - cosine_precision@10 - cosine_recall@1 - cosine_recall@3 - cosine_recall@5 - cosine_recall@10 - cosine_ndcg@10 - cosine_mrr@10 - cosine_map@100 - dot_accuracy@1 - dot_accuracy@3 - dot_accuracy@5 - dot_accuracy@10 - dot_precision@1 - dot_precision@3 - dot_precision@5 - dot_precision@10 - dot_recall@1 - dot_recall@3 - dot_recall@5 - dot_recall@10 - dot_ndcg@10 - dot_mrr@10 - dot_map@100 pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:714 - loss:MatryoshkaLoss - loss:MultipleNegativesRankingLoss widget: - source_sentence: What was revealed when an explanation of the system was demanded? sentences: - "Blueprint for an AI Bill of Rights. \nPanel Discussions to Inform the Blueprint\ \ for An AI Bill of Rights \nOSTP co-hosted a series of six panel discussions\ \ in collaboration with the Center for American Progress, \nthe Joint Center for\ \ Political and Economic Studies, New America, the German Marshall Fund, the Electronic\ \ \nPrivacy Information Center, and the Mozilla Foundation. The purpose of these\ \ convenings – recordings of \nwhich are publicly available online112 – was to\ \ bring together a variety of experts, practitioners, advocates \nand federal\ \ government officials to offer insights and analysis on the risks, harms, benefits,\ \ and \npolicy opportunities of automated systems. Each panel discussion was organized\ \ around a wide-ranging \ntheme, exploring current challenges and concerns and\ \ considering what an automated society that \nrespects democratic values should\ \ look like. These discussions focused on the topics of consumer" - "associated with misinformation or manipulation. \nInformation Integrity \nMS-1.1-003\ \ \nDisaggregate evaluation metrics by demographic factors to identify any \n\ discrepancies in how content provenance mechanisms work across diverse \npopulations.\ \ \nInformation Integrity; Harmful \nBias and Homogenization \nMS-1.1-004 Develop\ \ a suite of metrics to evaluate structured public feedback exercises \ninformed\ \ by representative AI Actors. \nHuman-AI Configuration; Harmful \nBias and Homogenization;\ \ CBRN \nInformation or Capabilities \nMS-1.1-005 \nEvaluate novel methods and\ \ technologies for the measurement of GAI-related \nrisks including in content\ \ provenance, offensive cyber, and CBRN, while \nmaintaining the models’ ability\ \ to produce valid, reliable, and factually accurate \noutputs. \nInformation\ \ Integrity; CBRN \nInformation or Capabilities; \nObscene, Degrading, and/or\ \ \nAbusive Content" - 'errors and other system flaws. These flaws were only revealed when an explanation of the system was demanded and produced.86 The lack of an explanation made it harder for errors to be corrected in a timely manner. 42' - source_sentence: What should an entity do if an automated system leads to different treatment of identified groups? sentences: - "instance where the deployed automated system leads to different treatment or\ \ impacts disfavoring the identi­\nfied groups, the entity governing, implementing,\ \ or using the system should document the disparity and a \njustification for\ \ any continued use of the system. \nDisparity mitigation. When a disparity assessment\ \ identifies a disparity against an assessed group, it may \nbe appropriate to\ \ take steps to mitigate or eliminate the disparity. In some cases, mitigation\ \ or elimination of \nthe disparity may be required by law. \nDisparities that\ \ have the potential to lead to algorithmic \ndiscrimination, cause meaningful\ \ harm, or violate equity49 goals should be mitigated. When designing and \nevaluating\ \ an automated system, steps should be taken to evaluate multiple models and select\ \ the one that \nhas the least adverse impact, modify data input choices, or otherwise\ \ identify a system with fewer" - "48 \n• Data protection \n• Data retention \n• Consistency in use of defining\ \ key terms \n• Decommissioning \n• Discouraging anonymous use \n• Education \ \ \n• Impact assessments \n• Incident response \n• Monitoring \n• Opt-outs \n\ • Risk-based controls \n• Risk mapping and measurement \n• Science-backed TEVV\ \ practices \n• Secure software development practices \n• Stakeholder engagement\ \ \n• Synthetic content detection and \nlabeling tools and techniques \n• Whistleblower\ \ protections \n• Workforce diversity and \ninterdisciplinary teams\nEstablishing\ \ acceptable use policies and guidance for the use of GAI in formal human-AI teaming\ \ settings \nas well as different levels of human-AI configurations can help to\ \ decrease risks arising from misuse, \nabuse, inappropriate repurpose, and misalignment\ \ between systems and users. These practices are just \none example of adapting\ \ existing governance protocols for GAI contexts. \nA.1.3. Third-Party Considerations" - "in the spreading and scaling of harms. Data from some domains, including criminal\ \ justice data and data indi­\ncating adverse outcomes in domains such as finance,\ \ employment, and housing, is especially sensitive, and in \nsome cases its reuse\ \ is limited by law. Accordingly, such data should be subject to extra oversight\ \ to ensure \nsafety and efficacy. Data reuse of sensitive domain data in other\ \ contexts (e.g., criminal data reuse for civil legal \nmatters or private sector\ \ use) should only occur where use of such data is legally authorized and, after\ \ examina­\ntion, has benefits for those impacted by the system that outweigh\ \ identified risks and, as appropriate, reason­\nable measures have been implemented\ \ to mitigate the identified risks. Such data should be clearly labeled to \n\ identify contexts for limited reuse based on sensitivity. Where possible, aggregated\ \ datasets may be useful for \nreplacing individual-level sensitive data. \nDemonstrate\ \ the safety and effectiveness of the system" - source_sentence: What types of data are considered sensitive for individuals who are not yet legal adults? sentences: - "SAFE AND EFFECTIVE \nSYSTEMS \nWHY THIS PRINCIPLE IS IMPORTANT\nThis section\ \ provides a brief summary of the problems which the principle seeks to address\ \ and protect \nagainst, including illustrative examples. \n•\nAI-enabled “nudification”\ \ technology that creates images where people appear to be nude—including apps\ \ that\nenable non-technical users to create or alter images of individuals without\ \ their consent—has proliferated at an\nalarming rate. Such technology is becoming\ \ a common form of image-based abuse that disproportionately\nimpacts women. As\ \ these tools become more sophisticated, they are producing altered images that\ \ are increasing­\nly realistic and are difficult for both humans and AI to detect\ \ as inauthentic. Regardless of authenticity, the expe­\nrience of harm to victims\ \ of non-consensual intimate images can be devastatingly real—affecting their\ \ personal\nand professional lives, and impacting their mental and physical health.10\n\ •" - "those who are not yet legal adults is also sensitive, even if not related to\ \ a sensitive domain. Such data includes, \nbut is not limited to, numerical,\ \ text, image, audio, or video data. \nSENSITIVE DOMAINS: “Sensitive domains”\ \ are those in which activities being conducted can cause material \nharms, including\ \ significant adverse effects on human rights such as autonomy and dignity, as\ \ well as civil liber­\nties and civil rights. Domains that have historically\ \ been singled out as deserving of enhanced data protections \nor where such enhanced\ \ protections are reasonably expected by the public include, but are not limited\ \ to, \nhealth, family planning and care, employment, education, criminal justice,\ \ and personal finance. In the context \nof this framework, such domains are considered\ \ sensitive whether or not the specifics of a system context \nwould necessitate\ \ coverage under existing law, and domains and data that are considered sensitive\ \ are under­" - "Homogenization; Intellectual \nProperty \nAI Actor Tasks: Governance and Oversight\ \ \n \n \n \n14 AI Actors are defined by the OECD as “those who play an active\ \ role in the AI system lifecycle, including \norganizations and individuals that\ \ deploy or operate AI.” See Appendix A of the AI RMF for additional descriptions\ \ \nof AI Actors and AI Actor Tasks." - source_sentence: What are the main findings of Carlini et al. (2021) regarding training data extraction from large language models? sentences: - "59 \nTirrell, L. (2017) Toxic Speech: Toward an Epidemiology of Discursive Harm.\ \ Philosophical Topics, 45(2), \n139-162. https://www.jstor.org/stable/26529441\ \ \nTufekci, Z. (2015) Algorithmic Harms Beyond Facebook and Google: Emergent\ \ Challenges of \nComputational Agency. Colorado Technology Law Journal. https://ctlj.colorado.edu/wp-\n\ content/uploads/2015/08/Tufekci-final.pdf \nTurri, V. et al. (2023) Why We Need\ \ to Know More: Exploring the State of AI Incident Documentation \nPractices.\ \ AAAI/ACM Conference on AI, Ethics, and Society. \nhttps://dl.acm.org/doi/fullHtml/10.1145/3600211.3604700\ \ \nUrbina, F. et al. (2022) Dual use of artificial-intelligence-powered drug discovery.\ \ Nature Machine \nIntelligence. https://www.nature.com/articles/s42256-022-00465-9\ \ \nWang, X. et al. (2023) Energy and Carbon Considerations of Fine-Tuning BERT.\ \ ACL Anthology. \nhttps://aclanthology.org/2023.findings-emnlp.607.pdf \nWang,\ \ Y. et al. (2023) Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs.\ \ arXiv." - "8 \nTrustworthy AI Characteristics: Accountable and Transparent, Privacy Enhanced,\ \ Safe, Secure and \nResilient \n2.5. Environmental Impacts \nTraining, maintaining,\ \ and operating (running inference on) GAI systems are resource-intensive activities,\ \ \nwith potentially large energy and environmental footprints. Energy and carbon\ \ emissions vary based on \nwhat is being done with the GAI model (i.e., pre-training,\ \ fine-tuning, inference), the modality of the \ncontent, hardware used, and type\ \ of task or application. \nCurrent estimates suggest that training a single transformer\ \ LLM can emit as much carbon as 300 round-\ntrip flights between San Francisco\ \ and New York. In a study comparing energy consumption and carbon \nemissions\ \ for LLM inference, generative tasks (e.g., text summarization) were found to\ \ be more energy- \nand carbon-intensive than discriminative or non-generative\ \ tasks (e.g., text classification)." - "Carlini, N., et al. (2021) Extracting Training Data from Large Language Models.\ \ Usenix. \nhttps://www.usenix.org/conference/usenixsecurity21/presentation/carlini-extracting\ \ \nCarlini, N. et al. (2023) Quantifying Memorization Across Neural Language\ \ Models. ICLR 2023. \nhttps://arxiv.org/pdf/2202.07646 \nCarlini, N. et al. (2024)\ \ Stealing Part of a Production Language Model. arXiv. \nhttps://arxiv.org/abs/2403.06634\ \ \nChandra, B. et al. (2023) Dismantling the Disinformation Business of Chinese\ \ Influence Operations. \nRAND. https://www.rand.org/pubs/commentary/2023/10/dismantling-the-disinformation-business-of-\n\ chinese.html \nCiriello, R. et al. (2024) Ethical Tensions in Human-AI Companionship:\ \ A Dialectical Inquiry into Replika. \nResearchGate. https://www.researchgate.net/publication/374505266_Ethical_Tensions_in_Human-\n\ AI_Companionship_A_Dialectical_Inquiry_into_Replika \nDahl, M. et al. (2024) Large\ \ Legal Fictions: Profiling Legal Hallucinations in Large Language Models. arXiv." - source_sentence: What are the potential risks associated with confabulated outputs in healthcare applications? sentences: - "glossary of terms pertinent to GAI risk management will be developed and hosted\ \ on NIST’s Trustworthy & \nResponsible AI Resource Center (AIRC), and added to\ \ The Language of Trustworthy AI: An In-Depth Glossary of \nTerms. \nThis document\ \ was also informed by public comments and consultations from several Requests\ \ for Information. \n \n2. \nOverview of Risks Unique to or Exacerbated by GAI\ \ \nIn the context of the AI RMF, risk refers to the composite measure of an event’s\ \ probability (or \nlikelihood) of occurring and the magnitude or degree of the\ \ consequences of the corresponding event. \nSome risks can be assessed as likely\ \ to materialize in a given context, particularly those that have been \nempirically\ \ demonstrated in similar contexts. Other risks may be unlikely to materialize\ \ in a given \ncontext, or may be more speculative and therefore uncertain. \n\ AI risks can differ from or intensify traditional software risks. Likewise, GAI\ \ can exacerbate existing AI" - "outputs that are factually inaccurate or internally inconsistent. This dynamic\ \ is particularly relevant when \nit comes to open-ended prompts for long-form\ \ responses and in domains which require highly \ncontextual and/or domain expertise.\ \ \nRisks from confabulations may arise when users believe false content – often\ \ due to the confident nature \nof the response – leading users to act upon or\ \ promote the false information. This poses a challenge for \nmany real-world\ \ applications, such as in healthcare, where a confabulated summary of patient\ \ \ninformation reports could cause doctors to make incorrect diagnoses and/or\ \ recommend the wrong \ntreatments. Risks of confabulated content may be especially\ \ important to monitor when integrating GAI \ninto applications involving consequential\ \ decision making. \nGAI outputs may also include confabulated logic or citations\ \ that purport to justify or explain the" - "ABOUT THIS FRAMEWORK­­­­­\nThe Blueprint for an AI Bill of Rights is a set of\ \ five principles and associated practices to help guide the \ndesign, use, and\ \ deployment of automated systems to protect the rights of the American public\ \ in the age of \nartificial intel-ligence. Developed through extensive consultation\ \ with the American public, these principles are \na blueprint for building and\ \ deploying automated systems that are aligned with democratic values and protect\ \ \ncivil rights, civil liberties, and privacy. The Blueprint for an AI Bill of\ \ Rights includes this Foreword, the five \nprinciples, notes on Applying the\ \ The Blueprint for an AI Bill of Rights, and a Technical Companion that gives\ \ \nconcrete steps that can be taken by many kinds of organizations—from governments\ \ at all levels to companies of \nall sizes—to uphold these values. Experts from\ \ across the private sector, governments, and international" model-index: - name: SentenceTransformer based on Snowflake/snowflake-arctic-embed-m results: - task: type: information-retrieval name: Information Retrieval dataset: name: Unknown type: unknown metrics: - type: cosine_accuracy@1 value: 0.881578947368421 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.9671052631578947 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.9868421052631579 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 1.0 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.881578947368421 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.3223684210526316 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.19736842105263155 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.09999999999999999 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.881578947368421 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.9671052631578947 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.9868421052631579 name: Cosine Recall@5 - type: cosine_recall@10 value: 1.0 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9432705827144577 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9245614035087719 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9245614035087719 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.881578947368421 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.9671052631578947 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 0.9868421052631579 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 1.0 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.881578947368421 name: Dot Precision@1 - type: dot_precision@3 value: 0.3223684210526316 name: Dot Precision@3 - type: dot_precision@5 value: 0.19736842105263155 name: Dot Precision@5 - type: dot_precision@10 value: 0.09999999999999999 name: Dot Precision@10 - type: dot_recall@1 value: 0.881578947368421 name: Dot Recall@1 - type: dot_recall@3 value: 0.9671052631578947 name: Dot Recall@3 - type: dot_recall@5 value: 0.9868421052631579 name: Dot Recall@5 - type: dot_recall@10 value: 1.0 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9432705827144577 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9245614035087719 name: Dot Mrr@10 - type: dot_map@100 value: 0.9245614035087719 name: Dot Map@100 --- # SentenceTransformer based on Snowflake/snowflake-arctic-embed-m This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m) <!-- at revision e2b128b9fa60c82b4585512b33e1544224ffff42 --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 768 tokens - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("jet-taekyo/snowflake_finetuned_recursive") # Run inference sentences = [ 'What are the potential risks associated with confabulated outputs in healthcare applications?', 'outputs that are factually inaccurate or internally inconsistent. This dynamic is particularly relevant when \nit comes to open-ended prompts for long-form responses and in domains which require highly \ncontextual and/or domain expertise. \nRisks from confabulations may arise when users believe false content – often due to the confident nature \nof the response – leading users to act upon or promote the false information. This poses a challenge for \nmany real-world applications, such as in healthcare, where a confabulated summary of patient \ninformation reports could cause doctors to make incorrect diagnoses and/or recommend the wrong \ntreatments. Risks of confabulated content may be especially important to monitor when integrating GAI \ninto applications involving consequential decision making. \nGAI outputs may also include confabulated logic or citations that purport to justify or explain the', 'ABOUT THIS FRAMEWORK\xad\xad\xad\xad\xad\nThe Blueprint for an AI Bill of Rights is a set of five principles and associated practices to help guide the \ndesign, use, and deployment of automated systems to protect the rights of the American public in the age of \nartificial intel-ligence. Developed through extensive consultation with the American public, these principles are \na blueprint for building and deploying automated systems that are aligned with democratic values and protect \ncivil rights, civil liberties, and privacy. The Blueprint for an AI Bill of Rights includes this Foreword, the five \nprinciples, notes on Applying the The Blueprint for an AI Bill of Rights, and a Technical Companion that gives \nconcrete steps that can be taken by many kinds of organizations—from governments at all levels to companies of \nall sizes—to uphold these values. Experts from across the private sector, governments, and international', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:-----------| | cosine_accuracy@1 | 0.8816 | | cosine_accuracy@3 | 0.9671 | | cosine_accuracy@5 | 0.9868 | | cosine_accuracy@10 | 1.0 | | cosine_precision@1 | 0.8816 | | cosine_precision@3 | 0.3224 | | cosine_precision@5 | 0.1974 | | cosine_precision@10 | 0.1 | | cosine_recall@1 | 0.8816 | | cosine_recall@3 | 0.9671 | | cosine_recall@5 | 0.9868 | | cosine_recall@10 | 1.0 | | cosine_ndcg@10 | 0.9433 | | cosine_mrr@10 | 0.9246 | | **cosine_map@100** | **0.9246** | | dot_accuracy@1 | 0.8816 | | dot_accuracy@3 | 0.9671 | | dot_accuracy@5 | 0.9868 | | dot_accuracy@10 | 1.0 | | dot_precision@1 | 0.8816 | | dot_precision@3 | 0.3224 | | dot_precision@5 | 0.1974 | | dot_precision@10 | 0.1 | | dot_recall@1 | 0.8816 | | dot_recall@3 | 0.9671 | | dot_recall@5 | 0.9868 | | dot_recall@10 | 1.0 | | dot_ndcg@10 | 0.9433 | | dot_mrr@10 | 0.9246 | | dot_map@100 | 0.9246 | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 714 training samples * Columns: <code>sentence_0</code> and <code>sentence_1</code> * Approximate statistics based on the first 714 samples: | | sentence_0 | sentence_1 | |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| | type | string | string | | details | <ul><li>min: 10 tokens</li><li>mean: 18.5 tokens</li><li>max: 36 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 178.94 tokens</li><li>max: 512 tokens</li></ul> | * Samples: | sentence_0 | sentence_1 | |:------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>What are the implications of surveillance programs on contract lawyers' work?</code> | <code>https://humanimpact.org/wp-content/uploads/2021/01/The-Public-Health-Crisis-Hidden-In-Amazon­<br>Warehouses-HIP-WWRC-01-21.pdf; Drew Harwell. Contract lawyers face a growing invasion of<br>surveillance programs that monitor their work. The Washington Post. Nov. 11, 2021. https://<br>www.washingtonpost.com/technology/2021/11/11/lawyer-facial-recognition-monitoring/;<br>Virginia Doellgast and Sean O'Brady. Making Call Center Jobs Better: The Relationship between<br>Management Practices and Worker Stress. A Report for the CWA. June 2020. https://<br>hdl.handle.net/1813/74307<br>62. See, e.g., Federal Trade Commission. Data Brokers: A Call for Transparency and Accountability. May<br>2014.<br>https://www.ftc.gov/system/files/documents/reports/data-brokers-call-transparency-accountability­<br>report-federal-trade-commission-may-2014/140527databrokerreport.pdf; Cathy O’Neil.<br>Weapons of Math Destruction. Penguin Books. 2017.<br>https://en.wikipedia.org/wiki/Weapons_of_Math_Destruction</code> | | <code>How do management practices in call centers affect worker stress?</code> | <code>https://humanimpact.org/wp-content/uploads/2021/01/The-Public-Health-Crisis-Hidden-In-Amazon­<br>Warehouses-HIP-WWRC-01-21.pdf; Drew Harwell. Contract lawyers face a growing invasion of<br>surveillance programs that monitor their work. The Washington Post. Nov. 11, 2021. https://<br>www.washingtonpost.com/technology/2021/11/11/lawyer-facial-recognition-monitoring/;<br>Virginia Doellgast and Sean O'Brady. Making Call Center Jobs Better: The Relationship between<br>Management Practices and Worker Stress. A Report for the CWA. June 2020. https://<br>hdl.handle.net/1813/74307<br>62. See, e.g., Federal Trade Commission. Data Brokers: A Call for Transparency and Accountability. May<br>2014.<br>https://www.ftc.gov/system/files/documents/reports/data-brokers-call-transparency-accountability­<br>report-federal-trade-commission-may-2014/140527databrokerreport.pdf; Cathy O’Neil.<br>Weapons of Math Destruction. Penguin Books. 2017.<br>https://en.wikipedia.org/wiki/Weapons_of_Math_Destruction</code> | | <code>What major data breach involved the exposure of 235 million social media profiles?</code> | <code>65. See, e.g., Scott Ikeda. Major Data Broker Exposes 235 Million Social Media Profiles in Data Lead: Info<br>Appears to Have Been Scraped Without Permission. CPO Magazine. Aug. 28, 2020. https://<br>www.cpomagazine.com/cyber-security/major-data-broker-exposes-235-million-social-media-profiles­<br>in-data-leak/; Lily Hay Newman. 1.2 Billion Records Found Exposed Online in a Single Server. WIRED,<br>Nov. 22, 2019. https://www.wired.com/story/billion-records-exposed-online/<br>66. Lola Fadulu. Facial Recognition Technology in Public Housing Prompts Backlash. New York Times.<br>Sept. 24, 2019.<br>https://www.nytimes.com/2019/09/24/us/politics/facial-recognition-technology-housing.html<br>67. Jo Constantz. ‘They Were Spying On Us’: Amazon, Walmart, Use Surveillance Technology to Bust<br>Unions. Newsweek. Dec. 13, 2021.<br>https://www.newsweek.com/they-were-spying-us-amazon-walmart-use-surveillance-technology-bust­<br>unions-1658603<br>68. See, e.g., enforcement actions by the FTC against the photo storage app Everalbaum</code> | * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters: ```json { "loss": "MultipleNegativesRankingLoss", "matryoshka_dims": [ 768, 512, 256, 128, 64 ], "matryoshka_weights": [ 1, 1, 1, 1, 1 ], "n_dims_per_step": -1 } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `num_train_epochs`: 5 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 5 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `eval_use_gather_object`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | cosine_map@100 | |:------:|:----:|:--------------:| | 1.0 | 36 | 0.9149 | | 1.3889 | 50 | 0.9243 | | 2.0 | 72 | 0.9246 | ### Framework Versions - Python: 3.11.9 - Sentence Transformers: 3.1.0 - Transformers: 4.44.2 - PyTorch: 2.4.1+cu121 - Accelerate: 0.34.2 - Datasets: 3.0.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MatryoshkaLoss ```bibtex @misc{kusupati2024matryoshka, title={Matryoshka Representation Learning}, author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi}, year={2024}, eprint={2205.13147}, archivePrefix={arXiv}, primaryClass={cs.LG} } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
tom-brady/edge5
tom-brady
"2024-09-24T23:12:12Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:09:20Z"
Entry not found
jet-taekyo/snowflake_finetuned_semantic
jet-taekyo
"2024-09-24T23:09:50Z"
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:714", "loss:MatryoshkaLoss", "loss:MultipleNegativesRankingLoss", "arxiv:1908.10084", "arxiv:2205.13147", "arxiv:1705.00652", "base_model:Snowflake/snowflake-arctic-embed-m", "base_model:finetune:Snowflake/snowflake-arctic-embed-m", "model-index", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
"2024-09-24T23:09:29Z"
--- base_model: Snowflake/snowflake-arctic-embed-m library_name: sentence-transformers metrics: - cosine_accuracy@1 - cosine_accuracy@3 - cosine_accuracy@5 - cosine_accuracy@10 - cosine_precision@1 - cosine_precision@3 - cosine_precision@5 - cosine_precision@10 - cosine_recall@1 - cosine_recall@3 - cosine_recall@5 - cosine_recall@10 - cosine_ndcg@10 - cosine_mrr@10 - cosine_map@100 - dot_accuracy@1 - dot_accuracy@3 - dot_accuracy@5 - dot_accuracy@10 - dot_precision@1 - dot_precision@3 - dot_precision@5 - dot_precision@10 - dot_recall@1 - dot_recall@3 - dot_recall@5 - dot_recall@10 - dot_ndcg@10 - dot_mrr@10 - dot_map@100 pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:714 - loss:MatryoshkaLoss - loss:MultipleNegativesRankingLoss widget: - source_sentence: What was revealed when an explanation of the system was demanded? sentences: - "Blueprint for an AI Bill of Rights. \nPanel Discussions to Inform the Blueprint\ \ for An AI Bill of Rights \nOSTP co-hosted a series of six panel discussions\ \ in collaboration with the Center for American Progress, \nthe Joint Center for\ \ Political and Economic Studies, New America, the German Marshall Fund, the Electronic\ \ \nPrivacy Information Center, and the Mozilla Foundation. The purpose of these\ \ convenings – recordings of \nwhich are publicly available online112 – was to\ \ bring together a variety of experts, practitioners, advocates \nand federal\ \ government officials to offer insights and analysis on the risks, harms, benefits,\ \ and \npolicy opportunities of automated systems. Each panel discussion was organized\ \ around a wide-ranging \ntheme, exploring current challenges and concerns and\ \ considering what an automated society that \nrespects democratic values should\ \ look like. These discussions focused on the topics of consumer" - "associated with misinformation or manipulation. \nInformation Integrity \nMS-1.1-003\ \ \nDisaggregate evaluation metrics by demographic factors to identify any \n\ discrepancies in how content provenance mechanisms work across diverse \npopulations.\ \ \nInformation Integrity; Harmful \nBias and Homogenization \nMS-1.1-004 Develop\ \ a suite of metrics to evaluate structured public feedback exercises \ninformed\ \ by representative AI Actors. \nHuman-AI Configuration; Harmful \nBias and Homogenization;\ \ CBRN \nInformation or Capabilities \nMS-1.1-005 \nEvaluate novel methods and\ \ technologies for the measurement of GAI-related \nrisks including in content\ \ provenance, offensive cyber, and CBRN, while \nmaintaining the models’ ability\ \ to produce valid, reliable, and factually accurate \noutputs. \nInformation\ \ Integrity; CBRN \nInformation or Capabilities; \nObscene, Degrading, and/or\ \ \nAbusive Content" - 'errors and other system flaws. These flaws were only revealed when an explanation of the system was demanded and produced.86 The lack of an explanation made it harder for errors to be corrected in a timely manner. 42' - source_sentence: What should an entity do if an automated system leads to different treatment of identified groups? sentences: - "instance where the deployed automated system leads to different treatment or\ \ impacts disfavoring the identi­\nfied groups, the entity governing, implementing,\ \ or using the system should document the disparity and a \njustification for\ \ any continued use of the system. \nDisparity mitigation. When a disparity assessment\ \ identifies a disparity against an assessed group, it may \nbe appropriate to\ \ take steps to mitigate or eliminate the disparity. In some cases, mitigation\ \ or elimination of \nthe disparity may be required by law. \nDisparities that\ \ have the potential to lead to algorithmic \ndiscrimination, cause meaningful\ \ harm, or violate equity49 goals should be mitigated. When designing and \nevaluating\ \ an automated system, steps should be taken to evaluate multiple models and select\ \ the one that \nhas the least adverse impact, modify data input choices, or otherwise\ \ identify a system with fewer" - "48 \n• Data protection \n• Data retention \n• Consistency in use of defining\ \ key terms \n• Decommissioning \n• Discouraging anonymous use \n• Education \ \ \n• Impact assessments \n• Incident response \n• Monitoring \n• Opt-outs \n\ • Risk-based controls \n• Risk mapping and measurement \n• Science-backed TEVV\ \ practices \n• Secure software development practices \n• Stakeholder engagement\ \ \n• Synthetic content detection and \nlabeling tools and techniques \n• Whistleblower\ \ protections \n• Workforce diversity and \ninterdisciplinary teams\nEstablishing\ \ acceptable use policies and guidance for the use of GAI in formal human-AI teaming\ \ settings \nas well as different levels of human-AI configurations can help to\ \ decrease risks arising from misuse, \nabuse, inappropriate repurpose, and misalignment\ \ between systems and users. These practices are just \none example of adapting\ \ existing governance protocols for GAI contexts. \nA.1.3. Third-Party Considerations" - "in the spreading and scaling of harms. Data from some domains, including criminal\ \ justice data and data indi­\ncating adverse outcomes in domains such as finance,\ \ employment, and housing, is especially sensitive, and in \nsome cases its reuse\ \ is limited by law. Accordingly, such data should be subject to extra oversight\ \ to ensure \nsafety and efficacy. Data reuse of sensitive domain data in other\ \ contexts (e.g., criminal data reuse for civil legal \nmatters or private sector\ \ use) should only occur where use of such data is legally authorized and, after\ \ examina­\ntion, has benefits for those impacted by the system that outweigh\ \ identified risks and, as appropriate, reason­\nable measures have been implemented\ \ to mitigate the identified risks. Such data should be clearly labeled to \n\ identify contexts for limited reuse based on sensitivity. Where possible, aggregated\ \ datasets may be useful for \nreplacing individual-level sensitive data. \nDemonstrate\ \ the safety and effectiveness of the system" - source_sentence: What types of data are considered sensitive for individuals who are not yet legal adults? sentences: - "SAFE AND EFFECTIVE \nSYSTEMS \nWHY THIS PRINCIPLE IS IMPORTANT\nThis section\ \ provides a brief summary of the problems which the principle seeks to address\ \ and protect \nagainst, including illustrative examples. \n•\nAI-enabled “nudification”\ \ technology that creates images where people appear to be nude—including apps\ \ that\nenable non-technical users to create or alter images of individuals without\ \ their consent—has proliferated at an\nalarming rate. Such technology is becoming\ \ a common form of image-based abuse that disproportionately\nimpacts women. As\ \ these tools become more sophisticated, they are producing altered images that\ \ are increasing­\nly realistic and are difficult for both humans and AI to detect\ \ as inauthentic. Regardless of authenticity, the expe­\nrience of harm to victims\ \ of non-consensual intimate images can be devastatingly real—affecting their\ \ personal\nand professional lives, and impacting their mental and physical health.10\n\ •" - "those who are not yet legal adults is also sensitive, even if not related to\ \ a sensitive domain. Such data includes, \nbut is not limited to, numerical,\ \ text, image, audio, or video data. \nSENSITIVE DOMAINS: “Sensitive domains”\ \ are those in which activities being conducted can cause material \nharms, including\ \ significant adverse effects on human rights such as autonomy and dignity, as\ \ well as civil liber­\nties and civil rights. Domains that have historically\ \ been singled out as deserving of enhanced data protections \nor where such enhanced\ \ protections are reasonably expected by the public include, but are not limited\ \ to, \nhealth, family planning and care, employment, education, criminal justice,\ \ and personal finance. In the context \nof this framework, such domains are considered\ \ sensitive whether or not the specifics of a system context \nwould necessitate\ \ coverage under existing law, and domains and data that are considered sensitive\ \ are under­" - "Homogenization; Intellectual \nProperty \nAI Actor Tasks: Governance and Oversight\ \ \n \n \n \n14 AI Actors are defined by the OECD as “those who play an active\ \ role in the AI system lifecycle, including \norganizations and individuals that\ \ deploy or operate AI.” See Appendix A of the AI RMF for additional descriptions\ \ \nof AI Actors and AI Actor Tasks." - source_sentence: What are the main findings of Carlini et al. (2021) regarding training data extraction from large language models? sentences: - "59 \nTirrell, L. (2017) Toxic Speech: Toward an Epidemiology of Discursive Harm.\ \ Philosophical Topics, 45(2), \n139-162. https://www.jstor.org/stable/26529441\ \ \nTufekci, Z. (2015) Algorithmic Harms Beyond Facebook and Google: Emergent\ \ Challenges of \nComputational Agency. Colorado Technology Law Journal. https://ctlj.colorado.edu/wp-\n\ content/uploads/2015/08/Tufekci-final.pdf \nTurri, V. et al. (2023) Why We Need\ \ to Know More: Exploring the State of AI Incident Documentation \nPractices.\ \ AAAI/ACM Conference on AI, Ethics, and Society. \nhttps://dl.acm.org/doi/fullHtml/10.1145/3600211.3604700\ \ \nUrbina, F. et al. (2022) Dual use of artificial-intelligence-powered drug discovery.\ \ Nature Machine \nIntelligence. https://www.nature.com/articles/s42256-022-00465-9\ \ \nWang, X. et al. (2023) Energy and Carbon Considerations of Fine-Tuning BERT.\ \ ACL Anthology. \nhttps://aclanthology.org/2023.findings-emnlp.607.pdf \nWang,\ \ Y. et al. (2023) Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs.\ \ arXiv." - "8 \nTrustworthy AI Characteristics: Accountable and Transparent, Privacy Enhanced,\ \ Safe, Secure and \nResilient \n2.5. Environmental Impacts \nTraining, maintaining,\ \ and operating (running inference on) GAI systems are resource-intensive activities,\ \ \nwith potentially large energy and environmental footprints. Energy and carbon\ \ emissions vary based on \nwhat is being done with the GAI model (i.e., pre-training,\ \ fine-tuning, inference), the modality of the \ncontent, hardware used, and type\ \ of task or application. \nCurrent estimates suggest that training a single transformer\ \ LLM can emit as much carbon as 300 round-\ntrip flights between San Francisco\ \ and New York. In a study comparing energy consumption and carbon \nemissions\ \ for LLM inference, generative tasks (e.g., text summarization) were found to\ \ be more energy- \nand carbon-intensive than discriminative or non-generative\ \ tasks (e.g., text classification)." - "Carlini, N., et al. (2021) Extracting Training Data from Large Language Models.\ \ Usenix. \nhttps://www.usenix.org/conference/usenixsecurity21/presentation/carlini-extracting\ \ \nCarlini, N. et al. (2023) Quantifying Memorization Across Neural Language\ \ Models. ICLR 2023. \nhttps://arxiv.org/pdf/2202.07646 \nCarlini, N. et al. (2024)\ \ Stealing Part of a Production Language Model. arXiv. \nhttps://arxiv.org/abs/2403.06634\ \ \nChandra, B. et al. (2023) Dismantling the Disinformation Business of Chinese\ \ Influence Operations. \nRAND. https://www.rand.org/pubs/commentary/2023/10/dismantling-the-disinformation-business-of-\n\ chinese.html \nCiriello, R. et al. (2024) Ethical Tensions in Human-AI Companionship:\ \ A Dialectical Inquiry into Replika. \nResearchGate. https://www.researchgate.net/publication/374505266_Ethical_Tensions_in_Human-\n\ AI_Companionship_A_Dialectical_Inquiry_into_Replika \nDahl, M. et al. (2024) Large\ \ Legal Fictions: Profiling Legal Hallucinations in Large Language Models. arXiv." - source_sentence: What are the potential risks associated with confabulated outputs in healthcare applications? sentences: - "glossary of terms pertinent to GAI risk management will be developed and hosted\ \ on NIST’s Trustworthy & \nResponsible AI Resource Center (AIRC), and added to\ \ The Language of Trustworthy AI: An In-Depth Glossary of \nTerms. \nThis document\ \ was also informed by public comments and consultations from several Requests\ \ for Information. \n \n2. \nOverview of Risks Unique to or Exacerbated by GAI\ \ \nIn the context of the AI RMF, risk refers to the composite measure of an event’s\ \ probability (or \nlikelihood) of occurring and the magnitude or degree of the\ \ consequences of the corresponding event. \nSome risks can be assessed as likely\ \ to materialize in a given context, particularly those that have been \nempirically\ \ demonstrated in similar contexts. Other risks may be unlikely to materialize\ \ in a given \ncontext, or may be more speculative and therefore uncertain. \n\ AI risks can differ from or intensify traditional software risks. Likewise, GAI\ \ can exacerbate existing AI" - "outputs that are factually inaccurate or internally inconsistent. This dynamic\ \ is particularly relevant when \nit comes to open-ended prompts for long-form\ \ responses and in domains which require highly \ncontextual and/or domain expertise.\ \ \nRisks from confabulations may arise when users believe false content – often\ \ due to the confident nature \nof the response – leading users to act upon or\ \ promote the false information. This poses a challenge for \nmany real-world\ \ applications, such as in healthcare, where a confabulated summary of patient\ \ \ninformation reports could cause doctors to make incorrect diagnoses and/or\ \ recommend the wrong \ntreatments. Risks of confabulated content may be especially\ \ important to monitor when integrating GAI \ninto applications involving consequential\ \ decision making. \nGAI outputs may also include confabulated logic or citations\ \ that purport to justify or explain the" - "ABOUT THIS FRAMEWORK­­­­­\nThe Blueprint for an AI Bill of Rights is a set of\ \ five principles and associated practices to help guide the \ndesign, use, and\ \ deployment of automated systems to protect the rights of the American public\ \ in the age of \nartificial intel-ligence. Developed through extensive consultation\ \ with the American public, these principles are \na blueprint for building and\ \ deploying automated systems that are aligned with democratic values and protect\ \ \ncivil rights, civil liberties, and privacy. The Blueprint for an AI Bill of\ \ Rights includes this Foreword, the five \nprinciples, notes on Applying the\ \ The Blueprint for an AI Bill of Rights, and a Technical Companion that gives\ \ \nconcrete steps that can be taken by many kinds of organizations—from governments\ \ at all levels to companies of \nall sizes—to uphold these values. Experts from\ \ across the private sector, governments, and international" model-index: - name: SentenceTransformer based on Snowflake/snowflake-arctic-embed-m results: - task: type: information-retrieval name: Information Retrieval dataset: name: Unknown type: unknown metrics: - type: cosine_accuracy@1 value: 0.875 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.9539473684210527 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.9868421052631579 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 1.0 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.875 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.31798245614035087 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.19736842105263155 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.09999999999999999 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.875 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.9539473684210527 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.9868421052631579 name: Cosine Recall@5 - type: cosine_recall@10 value: 1.0 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9415026124716055 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9222117794486215 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9222117794486215 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.875 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.9539473684210527 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 0.9868421052631579 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 1.0 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.875 name: Dot Precision@1 - type: dot_precision@3 value: 0.31798245614035087 name: Dot Precision@3 - type: dot_precision@5 value: 0.19736842105263155 name: Dot Precision@5 - type: dot_precision@10 value: 0.09999999999999999 name: Dot Precision@10 - type: dot_recall@1 value: 0.875 name: Dot Recall@1 - type: dot_recall@3 value: 0.9539473684210527 name: Dot Recall@3 - type: dot_recall@5 value: 0.9868421052631579 name: Dot Recall@5 - type: dot_recall@10 value: 1.0 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9415026124716055 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9222117794486215 name: Dot Mrr@10 - type: dot_map@100 value: 0.9222117794486215 name: Dot Map@100 - type: cosine_accuracy@1 value: 0.8984375 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.9765625 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.984375 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 1.0 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.8984375 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.32552083333333326 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.19687500000000002 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.10000000000000002 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.8984375 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.9765625 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.984375 name: Cosine Recall@5 - type: cosine_recall@10 value: 1.0 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9524988746459724 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9370008680555555 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9370008680555556 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.8984375 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.9765625 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 0.984375 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 1.0 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.8984375 name: Dot Precision@1 - type: dot_precision@3 value: 0.32552083333333326 name: Dot Precision@3 - type: dot_precision@5 value: 0.19687500000000002 name: Dot Precision@5 - type: dot_precision@10 value: 0.10000000000000002 name: Dot Precision@10 - type: dot_recall@1 value: 0.8984375 name: Dot Recall@1 - type: dot_recall@3 value: 0.9765625 name: Dot Recall@3 - type: dot_recall@5 value: 0.984375 name: Dot Recall@5 - type: dot_recall@10 value: 1.0 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9524988746459724 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9370008680555555 name: Dot Mrr@10 - type: dot_map@100 value: 0.9370008680555556 name: Dot Map@100 --- # SentenceTransformer based on Snowflake/snowflake-arctic-embed-m This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m) <!-- at revision e2b128b9fa60c82b4585512b33e1544224ffff42 --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 768 tokens - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("jet-taekyo/snowflake_finetuned_semantic") # Run inference sentences = [ 'What are the potential risks associated with confabulated outputs in healthcare applications?', 'outputs that are factually inaccurate or internally inconsistent. This dynamic is particularly relevant when \nit comes to open-ended prompts for long-form responses and in domains which require highly \ncontextual and/or domain expertise. \nRisks from confabulations may arise when users believe false content – often due to the confident nature \nof the response – leading users to act upon or promote the false information. This poses a challenge for \nmany real-world applications, such as in healthcare, where a confabulated summary of patient \ninformation reports could cause doctors to make incorrect diagnoses and/or recommend the wrong \ntreatments. Risks of confabulated content may be especially important to monitor when integrating GAI \ninto applications involving consequential decision making. \nGAI outputs may also include confabulated logic or citations that purport to justify or explain the', 'ABOUT THIS FRAMEWORK\xad\xad\xad\xad\xad\nThe Blueprint for an AI Bill of Rights is a set of five principles and associated practices to help guide the \ndesign, use, and deployment of automated systems to protect the rights of the American public in the age of \nartificial intel-ligence. Developed through extensive consultation with the American public, these principles are \na blueprint for building and deploying automated systems that are aligned with democratic values and protect \ncivil rights, civil liberties, and privacy. The Blueprint for an AI Bill of Rights includes this Foreword, the five \nprinciples, notes on Applying the The Blueprint for an AI Bill of Rights, and a Technical Companion that gives \nconcrete steps that can be taken by many kinds of organizations—from governments at all levels to companies of \nall sizes—to uphold these values. Experts from across the private sector, governments, and international', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:-----------| | cosine_accuracy@1 | 0.875 | | cosine_accuracy@3 | 0.9539 | | cosine_accuracy@5 | 0.9868 | | cosine_accuracy@10 | 1.0 | | cosine_precision@1 | 0.875 | | cosine_precision@3 | 0.318 | | cosine_precision@5 | 0.1974 | | cosine_precision@10 | 0.1 | | cosine_recall@1 | 0.875 | | cosine_recall@3 | 0.9539 | | cosine_recall@5 | 0.9868 | | cosine_recall@10 | 1.0 | | cosine_ndcg@10 | 0.9415 | | cosine_mrr@10 | 0.9222 | | **cosine_map@100** | **0.9222** | | dot_accuracy@1 | 0.875 | | dot_accuracy@3 | 0.9539 | | dot_accuracy@5 | 0.9868 | | dot_accuracy@10 | 1.0 | | dot_precision@1 | 0.875 | | dot_precision@3 | 0.318 | | dot_precision@5 | 0.1974 | | dot_precision@10 | 0.1 | | dot_recall@1 | 0.875 | | dot_recall@3 | 0.9539 | | dot_recall@5 | 0.9868 | | dot_recall@10 | 1.0 | | dot_ndcg@10 | 0.9415 | | dot_mrr@10 | 0.9222 | | dot_map@100 | 0.9222 | #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:----------| | cosine_accuracy@1 | 0.8984 | | cosine_accuracy@3 | 0.9766 | | cosine_accuracy@5 | 0.9844 | | cosine_accuracy@10 | 1.0 | | cosine_precision@1 | 0.8984 | | cosine_precision@3 | 0.3255 | | cosine_precision@5 | 0.1969 | | cosine_precision@10 | 0.1 | | cosine_recall@1 | 0.8984 | | cosine_recall@3 | 0.9766 | | cosine_recall@5 | 0.9844 | | cosine_recall@10 | 1.0 | | cosine_ndcg@10 | 0.9525 | | cosine_mrr@10 | 0.937 | | **cosine_map@100** | **0.937** | | dot_accuracy@1 | 0.8984 | | dot_accuracy@3 | 0.9766 | | dot_accuracy@5 | 0.9844 | | dot_accuracy@10 | 1.0 | | dot_precision@1 | 0.8984 | | dot_precision@3 | 0.3255 | | dot_precision@5 | 0.1969 | | dot_precision@10 | 0.1 | | dot_recall@1 | 0.8984 | | dot_recall@3 | 0.9766 | | dot_recall@5 | 0.9844 | | dot_recall@10 | 1.0 | | dot_ndcg@10 | 0.9525 | | dot_mrr@10 | 0.937 | | dot_map@100 | 0.937 | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 714 training samples * Columns: <code>sentence_0</code> and <code>sentence_1</code> * Approximate statistics based on the first 714 samples: | | sentence_0 | sentence_1 | |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| | type | string | string | | details | <ul><li>min: 7 tokens</li><li>mean: 17.22 tokens</li><li>max: 32 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 169.35 tokens</li><li>max: 512 tokens</li></ul> | * Samples: | sentence_0 | sentence_1 | |:------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>Who conducted the two listening sessions for members of the public?</code> | <code>APPENDIX<br>Lisa Feldman Barrett <br>Madeline Owens <br>Marsha Tudor <br>Microsoft Corporation <br>MITRE Corporation <br>National Association for the <br>Advancement of Colored People <br>Legal Defense and Educational <br>Fund <br>National Association of Criminal <br>Defense Lawyers <br>National Center for Missing & <br>Exploited Children <br>National Fair Housing Alliance <br>National Immigration Law Center <br>NEC Corporation of America <br>New America’s Open Technology <br>Institute <br>New York Civil Liberties Union <br>No Name Provided <br>Notre Dame Technology Ethics <br>Center <br>Office of the Ohio Public Defender <br>Onfido <br>Oosto <br>Orissa Rose <br>Palantir <br>Pangiam <br>Parity Technologies <br>Patrick A. Stewart, Jeffrey K. Mullins, and Thomas J. Greitens <br>Pel Abbott <br>Philadelphia Unemployment <br>Project <br>Project On Government Oversight <br>Recording Industry Association of <br>America <br>Robert Wilkens <br>Ron Hedges <br>Science, Technology, and Public <br>Policy Program at University of <br>Michigan Ann Arbor <br>Security Industry Association <br>Sheila Dean <br>Software & Information Industry <br>Association <br>Stephanie Dinkins and the Future <br>Histories Studio at Stony Brook <br>University <br>TechNet <br>The Alliance for Media Arts and <br>Culture, MIT Open Documentary <br>Lab and Co-Creation Studio, and <br>Immerse <br>The International Brotherhood of <br>Teamsters <br>The Leadership Conference on <br>Civil and Human Rights <br>Thorn <br>U.S. Chamber of Commerce’s <br>Technology Engagement Center <br>Uber Technologies <br>University of Pittsburgh <br>Undergraduate Student <br>Collaborative <br>Upturn <br>US Technology Policy Committee <br>of the Association of Computing <br>Machinery <br>Virginia Puccio <br>Visar Berisha and Julie Liss <br>XR Association <br>XR Safety Initiative <br>• As an additional effort to reach out to stakeholders regarding the RFI, OSTP conducted two listening sessions<br>for members of the public. The listening sessions together drew upwards of 300 participants.</code> | | <code>How many participants attended the listening sessions conducted by OSTP?</code> | <code>APPENDIX<br>Lisa Feldman Barrett <br>Madeline Owens <br>Marsha Tudor <br>Microsoft Corporation <br>MITRE Corporation <br>National Association for the <br>Advancement of Colored People <br>Legal Defense and Educational <br>Fund <br>National Association of Criminal <br>Defense Lawyers <br>National Center for Missing & <br>Exploited Children <br>National Fair Housing Alliance <br>National Immigration Law Center <br>NEC Corporation of America <br>New America’s Open Technology <br>Institute <br>New York Civil Liberties Union <br>No Name Provided <br>Notre Dame Technology Ethics <br>Center <br>Office of the Ohio Public Defender <br>Onfido <br>Oosto <br>Orissa Rose <br>Palantir <br>Pangiam <br>Parity Technologies <br>Patrick A. Stewart, Jeffrey K. Mullins, and Thomas J. Greitens <br>Pel Abbott <br>Philadelphia Unemployment <br>Project <br>Project On Government Oversight <br>Recording Industry Association of <br>America <br>Robert Wilkens <br>Ron Hedges <br>Science, Technology, and Public <br>Policy Program at University of <br>Michigan Ann Arbor <br>Security Industry Association <br>Sheila Dean <br>Software & Information Industry <br>Association <br>Stephanie Dinkins and the Future <br>Histories Studio at Stony Brook <br>University <br>TechNet <br>The Alliance for Media Arts and <br>Culture, MIT Open Documentary <br>Lab and Co-Creation Studio, and <br>Immerse <br>The International Brotherhood of <br>Teamsters <br>The Leadership Conference on <br>Civil and Human Rights <br>Thorn <br>U.S. Chamber of Commerce’s <br>Technology Engagement Center <br>Uber Technologies <br>University of Pittsburgh <br>Undergraduate Student <br>Collaborative <br>Upturn <br>US Technology Policy Committee <br>of the Association of Computing <br>Machinery <br>Virginia Puccio <br>Visar Berisha and Julie Liss <br>XR Association <br>XR Safety Initiative <br>• As an additional effort to reach out to stakeholders regarding the RFI, OSTP conducted two listening sessions<br>for members of the public. The listening sessions together drew upwards of 300 participants.</code> | | <code>What is the focus of the article from Wired regarding opioid drug addiction?</code> | <code>11,<br>2021. https://www.wired.com/story/opioid-drug-addiction-algorithm-chronic-pain/<br>104. Spencer Soper. Fired by Bot at Amazon: "It's You Against the Machine". Bloomberg, Jun. 28, 2021. https://www.bloomberg.com/news/features/2021-06-28/fired-by-bot-amazon-turns-to-machine­<br>managers-and-workers-are-losing-out<br>105. Definitions of ‘equity’ and ‘underserved communities’ can be found in the Definitions section of<br>this document as well as in Executive Order on Advancing Racial Equity and Support for Underserved<br>Communities Through the Federal Government:<br>https://www.whitehouse.gov/briefing-room/presidential-actions/2021/01/20/executive-order­<br>advancing-racial-equity-and-support-for-underserved-communities-through-the-federal-government/<br>106. HealthCare.gov.</code> | * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters: ```json { "loss": "MultipleNegativesRankingLoss", "matryoshka_dims": [ 768, 512, 256, 128, 64 ], "matryoshka_weights": [ 1, 1, 1, 1, 1 ], "n_dims_per_step": -1 } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `num_train_epochs`: 5 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 5 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `eval_use_gather_object`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | cosine_map@100 | |:------:|:----:|:--------------:| | 1.0 | 36 | 0.9149 | | 1.3889 | 50 | 0.9243 | | 2.0 | 72 | 0.9246 | | 2.7778 | 100 | 0.9217 | | 3.0 | 108 | 0.9211 | | 4.0 | 144 | 0.9233 | | 4.1667 | 150 | 0.9222 | | 5.0 | 180 | 0.9222 | | 1.0 | 31 | 0.9235 | | 1.6129 | 50 | 0.9365 | | 2.0 | 62 | 0.9347 | | 3.0 | 93 | 0.9350 | | 3.2258 | 100 | 0.9370 | ### Framework Versions - Python: 3.11.9 - Sentence Transformers: 3.1.0 - Transformers: 4.44.2 - PyTorch: 2.4.1+cu121 - Accelerate: 0.34.2 - Datasets: 3.0.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MatryoshkaLoss ```bibtex @misc{kusupati2024matryoshka, title={Matryoshka Representation Learning}, author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi}, year={2024}, eprint={2205.13147}, archivePrefix={arXiv}, primaryClass={cs.LG} } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
cc13qq/cifar10_wrn-28-10
cc13qq
"2024-09-24T23:10:29Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:09:30Z"
Entry not found
vapegod/bt19
vapegod
"2024-09-24T23:18:35Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:09:34Z"
Entry not found
SALUTEASD/Qwen-Qwen1.5-1.8B-1727219386
SALUTEASD
"2024-09-24T23:09:53Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:09:47Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
jet-taekyo/mpnet_finetuned_recursive
jet-taekyo
"2024-09-24T23:10:21Z"
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "mpnet", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:714", "loss:MatryoshkaLoss", "loss:MultipleNegativesRankingLoss", "arxiv:1908.10084", "arxiv:2205.13147", "arxiv:1705.00652", "base_model:sentence-transformers/all-mpnet-base-v2", "base_model:finetune:sentence-transformers/all-mpnet-base-v2", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
"2024-09-24T23:09:54Z"
--- base_model: sentence-transformers/all-mpnet-base-v2 library_name: sentence-transformers metrics: - cosine_accuracy@1 - cosine_accuracy@3 - cosine_accuracy@5 - cosine_accuracy@10 - cosine_precision@1 - cosine_precision@3 - cosine_precision@5 - cosine_precision@10 - cosine_recall@1 - cosine_recall@3 - cosine_recall@5 - cosine_recall@10 - cosine_ndcg@10 - cosine_mrr@10 - cosine_map@100 - dot_accuracy@1 - dot_accuracy@3 - dot_accuracy@5 - dot_accuracy@10 - dot_precision@1 - dot_precision@3 - dot_precision@5 - dot_precision@10 - dot_recall@1 - dot_recall@3 - dot_recall@5 - dot_recall@10 - dot_ndcg@10 - dot_mrr@10 - dot_map@100 pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:714 - loss:MatryoshkaLoss - loss:MultipleNegativesRankingLoss widget: - source_sentence: What is the purpose of comparing system performance with existing human performance after testing? sentences: - 'sensitive data without express parental consent, the lack of transparency in how such data is being used, and the potential for resulting discriminatory impacts. • Many employers transfer employee data to third party job verification services. This information is then used by potential future employers, banks, or landlords. In one case, a former employee alleged that a company supplied false data about her job title which resulted in a job offer being revoked.77 37' - "cate user choice or burden users with defaults that are privacy invasive. Con­\n\ sent should only be used to justify collection of data in cases where it can be\ \ \nappropriately and meaningfully given. Any consent requests should be brief,\ \ \nbe understandable in plain language, and give you agency over data collection\ \ \nand the specific context of use; current hard-to-understand no­\ntice-and-choice\ \ practices for broad uses of data should be changed. Enhanced \nprotections and\ \ restrictions for data and inferences related to sensitive do­\nmains, including\ \ health, work, education, criminal justice, and finance, and \nfor data pertaining\ \ to youth should put you first. In sensitive domains, your \ndata and related\ \ inferences should only be used for necessary functions, and \nyou should be\ \ protected by ethical review and use prohibitions. You and your \ncommunities\ \ should be free from unchecked surveillance; surveillance tech­\nnologies should\ \ be subject to heightened oversight that includes at least" - "systems testing and human-led (manual) testing. Testing conditions should mirror\ \ as closely as possible the \nconditions in which the system will be deployed,\ \ and new testing may be required for each deployment to \naccount for material\ \ differences in conditions from one deployment to another. Following testing,\ \ system \nperformance should be compared with the in-place, potentially human-driven,\ \ status quo procedures, with \nexisting human performance considered as a performance\ \ baseline for the algorithm to meet pre-deployment, \nand as a lifecycle minimum\ \ performance standard. Decision possibilities resulting from performance testing\ \ \nshould include the possibility of not deploying the system. \nRisk identification\ \ and mitigation. Before deployment, and in a proactive and ongoing manner, poten­\n\ tial risks of the automated system should be identified and mitigated. Identified\ \ risks should focus on the" - source_sentence: What steps should be taken to ensure automated systems are safe and effective before deployment? sentences: - "SAFE AND EFFECTIVE SYSTEMS \nYou should be protected from unsafe or ineffective\ \ sys­\ntems. Automated systems should be developed with consultation \nfrom diverse\ \ communities, stakeholders, and domain experts to iden­\ntify concerns, risks,\ \ and potential impacts of the system. Systems \nshould undergo pre-deployment\ \ testing, risk identification and miti­\ngation, and ongoing monitoring that\ \ demonstrate they are safe and \neffective based on their intended use, mitigation\ \ of unsafe outcomes \nincluding those beyond the intended use, and adherence\ \ to do­\nmain-specific standards. Outcomes of these protective measures \nshould\ \ include the possibility of not deploying the system or remov­\ning a system\ \ from use. Automated systems should not be designed \nwith an intent or reasonably\ \ foreseeable possibility of endangering \nyour safety or the safety of your community.\ \ They should be designed \nto proactively protect you from harms stemming from\ \ unintended," - 'ostp/2016_0504_data_discrimination.pdf; Cathy O’Neil. Weapons of Math Destruction. Penguin Books. 2017. https://en.wikipedia.org/wiki/Weapons_of_Math_Destruction; Ruha Benjamin. Race After Technology: Abolitionist Tools for the New Jim Code. Polity. 2019. https://www.ruhabenjamin.com/race­ after-technology 31. See, e.g., Kashmir Hill. Another Arrest, and Jail Time, Due to a Bad Facial Recognition Match: A New Jersey man was accused of shoplifting and trying to hit an officer with a car. He is the third known Black man to be wrongfully arrested based on face recognition. New York Times. Dec. 29, 2020, updated Jan. 6, 2021. https://www.nytimes.com/2020/12/29/technology/facial-recognition-misidentify-jail.html; Khari Johnson. How Wrongful Arrests Based on AI Derailed 3 Men''s Lives. Wired. Mar. 7, 2022. https:// www.wired.com/story/wrongful-arrests-ai-derailed-3-mens-lives/ 32. Student Borrower Protection Center. Educational Redlining. Student Borrower Protection Center' - "HUMAN ALTERNATIVES, \nCONSIDERATION, AND \nFALLBACK \nWHAT SHOULD BE EXPECTED\ \ OF AUTOMATED SYSTEMS\nThe expectations for automated systems are meant to serve\ \ as a blueprint for the development of additional \ntechnical standards and practices\ \ that are tailored for particular sectors and contexts. \nAn automated system\ \ should provide demonstrably effective mechanisms to opt out in favor of a human\ \ alterna­\ntive, where appropriate, as well as timely human consideration and\ \ remedy by a fallback system, with additional \nhuman oversight and safeguards\ \ for systems used in sensitive domains, and with training and assessment for\ \ any \nhuman-based portions of the system to ensure effectiveness. \nProvide\ \ a mechanism to conveniently opt out from automated systems in favor of a human\ \ \nalternative, where appropriate \nBrief, clear, accessible notice and instructions.\ \ Those impacted by an automated system should be" - source_sentence: How can high-integrity information be verified and authenticated? sentences: - "College of Medicine\n•\nMyaisha Hayes, Campaign Strategies Director, MediaJustice\n\ Panelists discussed uses of technology within the criminal justice system, including\ \ the use of predictive \npolicing, pretrial risk assessments, automated license\ \ plate readers, and prison communication tools. The \ndiscussion emphasized that\ \ communities deserve safety, and strategies need to be identified that lead to\ \ safety; \nsuch strategies might include data-driven approaches, but the focus\ \ on safety should be primary, and \ntechnology may or may not be part of an effective\ \ set of mechanisms to achieve safety. Various panelists raised \nconcerns about\ \ the validity of these systems, the tendency of adverse or irrelevant data to\ \ lead to a replication of \nunjust outcomes, and the confirmation bias and tendency\ \ of people to defer to potentially inaccurate automated \nsystems. Throughout,\ \ many of the panelists individually emphasized that the impact of these systems\ \ on" - "vetting. This information can be linked to the original source(s) with appropriate\ \ evidence. High-integrity \ninformation is also accurate and reliable, can be\ \ verified and authenticated, has a clear chain of custody, \nand creates reasonable\ \ expectations about when its validity may expire.”11 \n \n \n11 This definition\ \ of information integrity is derived from the 2022 White House Roadmap for Researchers\ \ on \nPriorities Related to Information Integrity Research and Development." - "consideration and fallback. In time-critical systems, this mechanism should be\ \ immediately available or, \nwhere possible, available before the harm occurs.\ \ Time-critical systems include, but are not limited to, \nvoting-related systems,\ \ automated building access and other access systems, systems that form a critical\ \ \ncomponent of healthcare, and systems that have the ability to withhold wages\ \ or otherwise cause \nimmediate financial penalties. \nEffective. The organizational\ \ structure surrounding processes for consideration and fallback should \nbe designed\ \ so that if the human decision-maker charged with reassessing a decision determines\ \ that it \nshould be overruled, the new decision will be effectively enacted.\ \ This includes ensuring that the new \ndecision is entered into the automated\ \ system throughout its components, any previous repercussions from \nthe old\ \ decision are also overturned, and safeguards are put in place to help ensure\ \ that future decisions do" - source_sentence: What concerns did panelists raise regarding access to broadband service in relation to healthcare delivery? sentences: - "systems, Incident response and containment. \nHuman-AI Configuration; \nInformation\ \ Security; Harmful Bias \nand Homogenization \nGV-3.2-003 \nDefine acceptable\ \ use policies for GAI interfaces, modalities, and human-AI \nconfigurations (i.e.,\ \ for chatbots and decision-making tasks), including criteria for \nthe kinds\ \ of queries GAI applications should refuse to respond to. \nHuman-AI Configuration\ \ \nGV-3.2-004 \nEstablish policies for user feedback mechanisms for GAI systems\ \ which include \nthorough instructions and any mechanisms for recourse. \nHuman-AI\ \ Configuration \nGV-3.2-005 \nEngage in threat modeling to anticipate potential\ \ risks from GAI systems. \nCBRN Information or Capabilities; \nInformation Security\ \ \nAI Actors: AI Design \n \nGOVERN 4.1: Organizational policies and practices\ \ are in place to foster a critical thinking and safety-first mindset in the design,\ \ \ndevelopment, deployment, and uses of AI systems to minimize potential negative\ \ impacts. \nAction ID \nSuggested Action \nGAI Risks" - "following items in GAI system inventory entries: Data provenance information\ \ \n(e.g., source, signatures, versioning, watermarks); Known issues reported\ \ from \ninternal bug tracking or external information sharing resources (e.g.,\ \ AI incident \ndatabase, AVID, CVE, NVD, or OECD AI incident monitor); Human\ \ oversight roles \nand responsibilities; Special rights and considerations for\ \ intellectual property, \nlicensed works, or personal, privileged, proprietary\ \ or sensitive data; Underlying \nfoundation models, versions of underlying models,\ \ and access modes. \nData Privacy; Human-AI \nConfiguration; Information \nIntegrity;\ \ Intellectual Property; \nValue Chain and Component \nIntegration \nAI Actor\ \ Tasks: Governance and Oversight" - "Sadie Tanner Mossell Alexander Professor of Civil Rights, University of Pennsylvania\n\ •\nDavid Jones, A. Bernard Ackerman Professor of the Culture of Medicine, Harvard\ \ University\n•\nJamila Michener, Associate Professor of Government, Cornell University;\ \ Co-Director, Cornell Center for\nHealth Equity­\nPanelists discussed the impact\ \ of new technologies on health disparities; healthcare access, delivery, and\ \ \noutcomes; and areas ripe for research and policymaking. Panelists discussed\ \ the increasing importance of tech-\nnology as both a vehicle to deliver healthcare\ \ and a tool to enhance the quality of care. On the issue of \ndelivery, various\ \ panelists pointed to a number of concerns including access to and expense of\ \ broadband \nservice, the privacy concerns associated with telehealth systems,\ \ the expense associated with health \nmonitoring devices, and how this can exacerbate\ \ equity issues. On the issue of technology enhanced care," - source_sentence: What is the purpose of the White House Office of Science and Technology Policy's initiative mentioned in the context? sentences: - "as ensuring that fallback mechanisms are in place to allow reversion to a previously\ \ working system. Monitor­\ning should take into account the performance of both\ \ technical system components (the algorithm as well as \nany hardware components,\ \ data inputs, etc.) and human operators. It should include mechanisms for testing\ \ \nthe actual accuracy of any predictions or recommendations generated by a system,\ \ not just a human operator’s \ndetermination of their accuracy. Ongoing monitoring\ \ procedures should include manual, human-led monitor­\ning as a check in the\ \ event there are shortcomings in automated monitoring systems. These monitoring\ \ proce­\ndures should be in place for the lifespan of the deployed automated\ \ system. \nClear organizational oversight. Entities responsible for the development\ \ or use of automated systems \nshould lay out clear governance structures and\ \ procedures. This includes clearly-stated governance proce­" - 'correct-signature-discrepancies.aspx 112. White House Office of Science and Technology Policy. Join the Effort to Create A Bill of Rights for an Automated Society. Nov. 10, 2021. https://www.whitehouse.gov/ostp/news-updates/2021/11/10/join-the-effort-to-create-a-bill-of­ rights-for-an-automated-society/ 113. White House Office of Science and Technology Policy. Notice of Request for Information (RFI) on Public and Private Sector Uses of Biometric Technologies. Issued Oct. 8, 2021. https://www.federalregister.gov/documents/2021/10/08/2021-21975/notice-of-request-for­ information-rfi-on-public-and-private-sector-uses-of-biometric-technologies 114. National Artificial Intelligence Initiative Office. Public Input on Public and Private Sector Uses of Biometric Technologies. Accessed Apr. 19, 2022. https://www.ai.gov/86-fr-56300-responses/ 115. Thomas D. Olszewski, Lisa M. Van Pay, Javier F. Ortiz, Sarah E. Swiersz, and Laurie A. Dacus.' - "Abusive Content \nMG-3.2-006 \nImplement real-time monitoring processes for analyzing\ \ generated content \nperformance and trustworthiness characteristics related\ \ to content provenance \nto identify deviations from the desired standards and\ \ trigger alerts for human \nintervention. \nInformation Integrity" model-index: - name: SentenceTransformer based on sentence-transformers/all-mpnet-base-v2 results: - task: type: information-retrieval name: Information Retrieval dataset: name: Unknown type: unknown metrics: - type: cosine_accuracy@1 value: 0.8618421052631579 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.9605263157894737 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.9868421052631579 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.993421052631579 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.8618421052631579 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.3201754385964913 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.19736842105263155 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.09934210526315788 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.8618421052631579 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.9605263157894737 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.9868421052631579 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.993421052631579 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9344707178079387 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9147556390977444 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9151668233082707 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.8618421052631579 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.9605263157894737 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 0.9868421052631579 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 0.993421052631579 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.8618421052631579 name: Dot Precision@1 - type: dot_precision@3 value: 0.3201754385964913 name: Dot Precision@3 - type: dot_precision@5 value: 0.19736842105263155 name: Dot Precision@5 - type: dot_precision@10 value: 0.09934210526315788 name: Dot Precision@10 - type: dot_recall@1 value: 0.8618421052631579 name: Dot Recall@1 - type: dot_recall@3 value: 0.9605263157894737 name: Dot Recall@3 - type: dot_recall@5 value: 0.9868421052631579 name: Dot Recall@5 - type: dot_recall@10 value: 0.993421052631579 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9344707178079387 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9147556390977444 name: Dot Mrr@10 - type: dot_map@100 value: 0.9151668233082707 name: Dot Map@100 --- # SentenceTransformer based on sentence-transformers/all-mpnet-base-v2 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) <!-- at revision 84f2bcc00d77236f9e89c8a360a00fb1139bf47d --> - **Maximum Sequence Length:** 384 tokens - **Output Dimensionality:** 768 tokens - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("jet-taekyo/mpnet_finetuned_recursive") # Run inference sentences = [ "What is the purpose of the White House Office of Science and Technology Policy's initiative mentioned in the context?", 'correct-signature-discrepancies.aspx\n112. White House Office of Science and Technology Policy. Join the Effort to Create A Bill of Rights for\nan Automated Society. Nov. 10, 2021.\nhttps://www.whitehouse.gov/ostp/news-updates/2021/11/10/join-the-effort-to-create-a-bill-of\xad\nrights-for-an-automated-society/\n113. White House Office of Science and Technology Policy. Notice of Request for Information (RFI) on\nPublic and Private Sector Uses of Biometric Technologies. Issued Oct. 8, 2021.\nhttps://www.federalregister.gov/documents/2021/10/08/2021-21975/notice-of-request-for\xad\ninformation-rfi-on-public-and-private-sector-uses-of-biometric-technologies\n114. National Artificial Intelligence Initiative Office. Public Input on Public and Private Sector Uses of\nBiometric Technologies. Accessed Apr. 19, 2022.\nhttps://www.ai.gov/86-fr-56300-responses/\n115. Thomas D. Olszewski, Lisa M. Van Pay, Javier F. Ortiz, Sarah E. Swiersz, and Laurie A. Dacus.', 'Abusive Content \nMG-3.2-006 \nImplement real-time monitoring processes for analyzing generated content \nperformance and trustworthiness characteristics related to content provenance \nto identify deviations from the desired standards and trigger alerts for human \nintervention. \nInformation Integrity', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:-----------| | cosine_accuracy@1 | 0.8618 | | cosine_accuracy@3 | 0.9605 | | cosine_accuracy@5 | 0.9868 | | cosine_accuracy@10 | 0.9934 | | cosine_precision@1 | 0.8618 | | cosine_precision@3 | 0.3202 | | cosine_precision@5 | 0.1974 | | cosine_precision@10 | 0.0993 | | cosine_recall@1 | 0.8618 | | cosine_recall@3 | 0.9605 | | cosine_recall@5 | 0.9868 | | cosine_recall@10 | 0.9934 | | cosine_ndcg@10 | 0.9345 | | cosine_mrr@10 | 0.9148 | | **cosine_map@100** | **0.9152** | | dot_accuracy@1 | 0.8618 | | dot_accuracy@3 | 0.9605 | | dot_accuracy@5 | 0.9868 | | dot_accuracy@10 | 0.9934 | | dot_precision@1 | 0.8618 | | dot_precision@3 | 0.3202 | | dot_precision@5 | 0.1974 | | dot_precision@10 | 0.0993 | | dot_recall@1 | 0.8618 | | dot_recall@3 | 0.9605 | | dot_recall@5 | 0.9868 | | dot_recall@10 | 0.9934 | | dot_ndcg@10 | 0.9345 | | dot_mrr@10 | 0.9148 | | dot_map@100 | 0.9152 | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 714 training samples * Columns: <code>sentence_0</code> and <code>sentence_1</code> * Approximate statistics based on the first 714 samples: | | sentence_0 | sentence_1 | |:--------|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| | type | string | string | | details | <ul><li>min: 10 tokens</li><li>mean: 18.48 tokens</li><li>max: 32 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 175.11 tokens</li><li>max: 384 tokens</li></ul> | * Samples: | sentence_0 | sentence_1 | |:----------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>What are the disadvantages faced by disabled students in virtual testing according to Heather Morrison?</code> | <code>74. See, e.g., Heather Morrison. Virtual Testing Puts Disabled Students at a Disadvantage. Government<br>Technology. May 24, 2022.<br>https://www.govtech.com/education/k-12/virtual-testing-puts-disabled-students-at-a-disadvantage;<br>Lydia X. Z. Brown, Ridhi Shetty, Matt Scherer, and Andrew Crawford. Ableism And Disability<br>Discrimination In New Surveillance Technologies: How new surveillance technologies in education,<br>policing, health care, and the workplace disproportionately harm disabled people. Center for Democracy<br>and Technology Report. May 24, 2022.<br>https://cdt.org/insights/ableism-and-disability-discrimination-in-new-surveillance-technologies-how­<br>new-surveillance-technologies-in-education-policing-health-care-and-the-workplace­<br>disproportionately-harm-disabled-people/<br>69</code> | | <code>How do new surveillance technologies disproportionately harm disabled people as discussed in the Center for Democracy and Technology Report?</code> | <code>74. See, e.g., Heather Morrison. Virtual Testing Puts Disabled Students at a Disadvantage. Government<br>Technology. May 24, 2022.<br>https://www.govtech.com/education/k-12/virtual-testing-puts-disabled-students-at-a-disadvantage;<br>Lydia X. Z. Brown, Ridhi Shetty, Matt Scherer, and Andrew Crawford. Ableism And Disability<br>Discrimination In New Surveillance Technologies: How new surveillance technologies in education,<br>policing, health care, and the workplace disproportionately harm disabled people. Center for Democracy<br>and Technology Report. May 24, 2022.<br>https://cdt.org/insights/ableism-and-disability-discrimination-in-new-surveillance-technologies-how­<br>new-surveillance-technologies-in-education-policing-health-care-and-the-workplace­<br>disproportionately-harm-disabled-people/<br>69</code> | | <code>What role does the National Highway Traffic Safety Administration play in ensuring vehicle safety?</code> | <code>The National Highway Traffic Safety Administration,14 through its rigorous standards and independent <br>evaluation, helps make sure vehicles on our roads are safe without limiting manufacturers’ ability to <br>innovate.15 At the same time, rules of the road are implemented locally to impose contextually appropriate <br>requirements on drivers, such as slowing down near schools or playgrounds.16<br>From large companies to start-ups, industry is providing innovative solutions that allow <br>organizations to mitigate risks to the safety and efficacy of AI systems, both before <br>deployment and through monitoring over time.17 These innovative solutions include risk <br>assessments, auditing mechanisms, assessment of organizational procedures, dashboards to allow for ongoing <br>monitoring, documentation procedures specific to model assessments, and many other strategies that aim to <br>mitigate risks posed by the use of AI to companies’ reputation, legal responsibilities, and other product safety</code> | * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters: ```json { "loss": "MultipleNegativesRankingLoss", "matryoshka_dims": [ 768, 512, 256, 128, 64 ], "matryoshka_weights": [ 1, 1, 1, 1, 1 ], "n_dims_per_step": -1 } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `num_train_epochs`: 5 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 5 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `eval_use_gather_object`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | cosine_map@100 | |:------:|:----:|:--------------:| | 1.0 | 36 | 0.9150 | | 1.3889 | 50 | 0.9152 | ### Framework Versions - Python: 3.11.9 - Sentence Transformers: 3.1.0 - Transformers: 4.44.2 - PyTorch: 2.4.1+cu121 - Accelerate: 0.34.2 - Datasets: 3.0.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MatryoshkaLoss ```bibtex @misc{kusupati2024matryoshka, title={Matryoshka Representation Learning}, author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi}, year={2024}, eprint={2205.13147}, archivePrefix={arXiv}, primaryClass={cs.LG} } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
jet-taekyo/mpnet_finetuned_semantic
jet-taekyo
"2024-09-24T23:10:41Z"
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "mpnet", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:714", "loss:MatryoshkaLoss", "loss:MultipleNegativesRankingLoss", "arxiv:1908.10084", "arxiv:2205.13147", "arxiv:1705.00652", "base_model:sentence-transformers/all-mpnet-base-v2", "base_model:finetune:sentence-transformers/all-mpnet-base-v2", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
"2024-09-24T23:10:23Z"
--- base_model: sentence-transformers/all-mpnet-base-v2 library_name: sentence-transformers metrics: - cosine_accuracy@1 - cosine_accuracy@3 - cosine_accuracy@5 - cosine_accuracy@10 - cosine_precision@1 - cosine_precision@3 - cosine_precision@5 - cosine_precision@10 - cosine_recall@1 - cosine_recall@3 - cosine_recall@5 - cosine_recall@10 - cosine_ndcg@10 - cosine_mrr@10 - cosine_map@100 - dot_accuracy@1 - dot_accuracy@3 - dot_accuracy@5 - dot_accuracy@10 - dot_precision@1 - dot_precision@3 - dot_precision@5 - dot_precision@10 - dot_recall@1 - dot_recall@3 - dot_recall@5 - dot_recall@10 - dot_ndcg@10 - dot_mrr@10 - dot_map@100 pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:714 - loss:MatryoshkaLoss - loss:MultipleNegativesRankingLoss widget: - source_sentence: What is the purpose of comparing system performance with existing human performance after testing? sentences: - 'sensitive data without express parental consent, the lack of transparency in how such data is being used, and the potential for resulting discriminatory impacts. • Many employers transfer employee data to third party job verification services. This information is then used by potential future employers, banks, or landlords. In one case, a former employee alleged that a company supplied false data about her job title which resulted in a job offer being revoked.77 37' - "cate user choice or burden users with defaults that are privacy invasive. Con­\n\ sent should only be used to justify collection of data in cases where it can be\ \ \nappropriately and meaningfully given. Any consent requests should be brief,\ \ \nbe understandable in plain language, and give you agency over data collection\ \ \nand the specific context of use; current hard-to-understand no­\ntice-and-choice\ \ practices for broad uses of data should be changed. Enhanced \nprotections and\ \ restrictions for data and inferences related to sensitive do­\nmains, including\ \ health, work, education, criminal justice, and finance, and \nfor data pertaining\ \ to youth should put you first. In sensitive domains, your \ndata and related\ \ inferences should only be used for necessary functions, and \nyou should be\ \ protected by ethical review and use prohibitions. You and your \ncommunities\ \ should be free from unchecked surveillance; surveillance tech­\nnologies should\ \ be subject to heightened oversight that includes at least" - "systems testing and human-led (manual) testing. Testing conditions should mirror\ \ as closely as possible the \nconditions in which the system will be deployed,\ \ and new testing may be required for each deployment to \naccount for material\ \ differences in conditions from one deployment to another. Following testing,\ \ system \nperformance should be compared with the in-place, potentially human-driven,\ \ status quo procedures, with \nexisting human performance considered as a performance\ \ baseline for the algorithm to meet pre-deployment, \nand as a lifecycle minimum\ \ performance standard. Decision possibilities resulting from performance testing\ \ \nshould include the possibility of not deploying the system. \nRisk identification\ \ and mitigation. Before deployment, and in a proactive and ongoing manner, poten­\n\ tial risks of the automated system should be identified and mitigated. Identified\ \ risks should focus on the" - source_sentence: What steps should be taken to ensure automated systems are safe and effective before deployment? sentences: - "SAFE AND EFFECTIVE SYSTEMS \nYou should be protected from unsafe or ineffective\ \ sys­\ntems. Automated systems should be developed with consultation \nfrom diverse\ \ communities, stakeholders, and domain experts to iden­\ntify concerns, risks,\ \ and potential impacts of the system. Systems \nshould undergo pre-deployment\ \ testing, risk identification and miti­\ngation, and ongoing monitoring that\ \ demonstrate they are safe and \neffective based on their intended use, mitigation\ \ of unsafe outcomes \nincluding those beyond the intended use, and adherence\ \ to do­\nmain-specific standards. Outcomes of these protective measures \nshould\ \ include the possibility of not deploying the system or remov­\ning a system\ \ from use. Automated systems should not be designed \nwith an intent or reasonably\ \ foreseeable possibility of endangering \nyour safety or the safety of your community.\ \ They should be designed \nto proactively protect you from harms stemming from\ \ unintended," - 'ostp/2016_0504_data_discrimination.pdf; Cathy O’Neil. Weapons of Math Destruction. Penguin Books. 2017. https://en.wikipedia.org/wiki/Weapons_of_Math_Destruction; Ruha Benjamin. Race After Technology: Abolitionist Tools for the New Jim Code. Polity. 2019. https://www.ruhabenjamin.com/race­ after-technology 31. See, e.g., Kashmir Hill. Another Arrest, and Jail Time, Due to a Bad Facial Recognition Match: A New Jersey man was accused of shoplifting and trying to hit an officer with a car. He is the third known Black man to be wrongfully arrested based on face recognition. New York Times. Dec. 29, 2020, updated Jan. 6, 2021. https://www.nytimes.com/2020/12/29/technology/facial-recognition-misidentify-jail.html; Khari Johnson. How Wrongful Arrests Based on AI Derailed 3 Men''s Lives. Wired. Mar. 7, 2022. https:// www.wired.com/story/wrongful-arrests-ai-derailed-3-mens-lives/ 32. Student Borrower Protection Center. Educational Redlining. Student Borrower Protection Center' - "HUMAN ALTERNATIVES, \nCONSIDERATION, AND \nFALLBACK \nWHAT SHOULD BE EXPECTED\ \ OF AUTOMATED SYSTEMS\nThe expectations for automated systems are meant to serve\ \ as a blueprint for the development of additional \ntechnical standards and practices\ \ that are tailored for particular sectors and contexts. \nAn automated system\ \ should provide demonstrably effective mechanisms to opt out in favor of a human\ \ alterna­\ntive, where appropriate, as well as timely human consideration and\ \ remedy by a fallback system, with additional \nhuman oversight and safeguards\ \ for systems used in sensitive domains, and with training and assessment for\ \ any \nhuman-based portions of the system to ensure effectiveness. \nProvide\ \ a mechanism to conveniently opt out from automated systems in favor of a human\ \ \nalternative, where appropriate \nBrief, clear, accessible notice and instructions.\ \ Those impacted by an automated system should be" - source_sentence: How can high-integrity information be verified and authenticated? sentences: - "College of Medicine\n•\nMyaisha Hayes, Campaign Strategies Director, MediaJustice\n\ Panelists discussed uses of technology within the criminal justice system, including\ \ the use of predictive \npolicing, pretrial risk assessments, automated license\ \ plate readers, and prison communication tools. The \ndiscussion emphasized that\ \ communities deserve safety, and strategies need to be identified that lead to\ \ safety; \nsuch strategies might include data-driven approaches, but the focus\ \ on safety should be primary, and \ntechnology may or may not be part of an effective\ \ set of mechanisms to achieve safety. Various panelists raised \nconcerns about\ \ the validity of these systems, the tendency of adverse or irrelevant data to\ \ lead to a replication of \nunjust outcomes, and the confirmation bias and tendency\ \ of people to defer to potentially inaccurate automated \nsystems. Throughout,\ \ many of the panelists individually emphasized that the impact of these systems\ \ on" - "vetting. This information can be linked to the original source(s) with appropriate\ \ evidence. High-integrity \ninformation is also accurate and reliable, can be\ \ verified and authenticated, has a clear chain of custody, \nand creates reasonable\ \ expectations about when its validity may expire.”11 \n \n \n11 This definition\ \ of information integrity is derived from the 2022 White House Roadmap for Researchers\ \ on \nPriorities Related to Information Integrity Research and Development." - "consideration and fallback. In time-critical systems, this mechanism should be\ \ immediately available or, \nwhere possible, available before the harm occurs.\ \ Time-critical systems include, but are not limited to, \nvoting-related systems,\ \ automated building access and other access systems, systems that form a critical\ \ \ncomponent of healthcare, and systems that have the ability to withhold wages\ \ or otherwise cause \nimmediate financial penalties. \nEffective. The organizational\ \ structure surrounding processes for consideration and fallback should \nbe designed\ \ so that if the human decision-maker charged with reassessing a decision determines\ \ that it \nshould be overruled, the new decision will be effectively enacted.\ \ This includes ensuring that the new \ndecision is entered into the automated\ \ system throughout its components, any previous repercussions from \nthe old\ \ decision are also overturned, and safeguards are put in place to help ensure\ \ that future decisions do" - source_sentence: What concerns did panelists raise regarding access to broadband service in relation to healthcare delivery? sentences: - "systems, Incident response and containment. \nHuman-AI Configuration; \nInformation\ \ Security; Harmful Bias \nand Homogenization \nGV-3.2-003 \nDefine acceptable\ \ use policies for GAI interfaces, modalities, and human-AI \nconfigurations (i.e.,\ \ for chatbots and decision-making tasks), including criteria for \nthe kinds\ \ of queries GAI applications should refuse to respond to. \nHuman-AI Configuration\ \ \nGV-3.2-004 \nEstablish policies for user feedback mechanisms for GAI systems\ \ which include \nthorough instructions and any mechanisms for recourse. \nHuman-AI\ \ Configuration \nGV-3.2-005 \nEngage in threat modeling to anticipate potential\ \ risks from GAI systems. \nCBRN Information or Capabilities; \nInformation Security\ \ \nAI Actors: AI Design \n \nGOVERN 4.1: Organizational policies and practices\ \ are in place to foster a critical thinking and safety-first mindset in the design,\ \ \ndevelopment, deployment, and uses of AI systems to minimize potential negative\ \ impacts. \nAction ID \nSuggested Action \nGAI Risks" - "following items in GAI system inventory entries: Data provenance information\ \ \n(e.g., source, signatures, versioning, watermarks); Known issues reported\ \ from \ninternal bug tracking or external information sharing resources (e.g.,\ \ AI incident \ndatabase, AVID, CVE, NVD, or OECD AI incident monitor); Human\ \ oversight roles \nand responsibilities; Special rights and considerations for\ \ intellectual property, \nlicensed works, or personal, privileged, proprietary\ \ or sensitive data; Underlying \nfoundation models, versions of underlying models,\ \ and access modes. \nData Privacy; Human-AI \nConfiguration; Information \nIntegrity;\ \ Intellectual Property; \nValue Chain and Component \nIntegration \nAI Actor\ \ Tasks: Governance and Oversight" - "Sadie Tanner Mossell Alexander Professor of Civil Rights, University of Pennsylvania\n\ •\nDavid Jones, A. Bernard Ackerman Professor of the Culture of Medicine, Harvard\ \ University\n•\nJamila Michener, Associate Professor of Government, Cornell University;\ \ Co-Director, Cornell Center for\nHealth Equity­\nPanelists discussed the impact\ \ of new technologies on health disparities; healthcare access, delivery, and\ \ \noutcomes; and areas ripe for research and policymaking. Panelists discussed\ \ the increasing importance of tech-\nnology as both a vehicle to deliver healthcare\ \ and a tool to enhance the quality of care. On the issue of \ndelivery, various\ \ panelists pointed to a number of concerns including access to and expense of\ \ broadband \nservice, the privacy concerns associated with telehealth systems,\ \ the expense associated with health \nmonitoring devices, and how this can exacerbate\ \ equity issues. On the issue of technology enhanced care," - source_sentence: What is the purpose of the White House Office of Science and Technology Policy's initiative mentioned in the context? sentences: - "as ensuring that fallback mechanisms are in place to allow reversion to a previously\ \ working system. Monitor­\ning should take into account the performance of both\ \ technical system components (the algorithm as well as \nany hardware components,\ \ data inputs, etc.) and human operators. It should include mechanisms for testing\ \ \nthe actual accuracy of any predictions or recommendations generated by a system,\ \ not just a human operator’s \ndetermination of their accuracy. Ongoing monitoring\ \ procedures should include manual, human-led monitor­\ning as a check in the\ \ event there are shortcomings in automated monitoring systems. These monitoring\ \ proce­\ndures should be in place for the lifespan of the deployed automated\ \ system. \nClear organizational oversight. Entities responsible for the development\ \ or use of automated systems \nshould lay out clear governance structures and\ \ procedures. This includes clearly-stated governance proce­" - 'correct-signature-discrepancies.aspx 112. White House Office of Science and Technology Policy. Join the Effort to Create A Bill of Rights for an Automated Society. Nov. 10, 2021. https://www.whitehouse.gov/ostp/news-updates/2021/11/10/join-the-effort-to-create-a-bill-of­ rights-for-an-automated-society/ 113. White House Office of Science and Technology Policy. Notice of Request for Information (RFI) on Public and Private Sector Uses of Biometric Technologies. Issued Oct. 8, 2021. https://www.federalregister.gov/documents/2021/10/08/2021-21975/notice-of-request-for­ information-rfi-on-public-and-private-sector-uses-of-biometric-technologies 114. National Artificial Intelligence Initiative Office. Public Input on Public and Private Sector Uses of Biometric Technologies. Accessed Apr. 19, 2022. https://www.ai.gov/86-fr-56300-responses/ 115. Thomas D. Olszewski, Lisa M. Van Pay, Javier F. Ortiz, Sarah E. Swiersz, and Laurie A. Dacus.' - "Abusive Content \nMG-3.2-006 \nImplement real-time monitoring processes for analyzing\ \ generated content \nperformance and trustworthiness characteristics related\ \ to content provenance \nto identify deviations from the desired standards and\ \ trigger alerts for human \nintervention. \nInformation Integrity" model-index: - name: SentenceTransformer based on sentence-transformers/all-mpnet-base-v2 results: - task: type: information-retrieval name: Information Retrieval dataset: name: Unknown type: unknown metrics: - type: cosine_accuracy@1 value: 0.8486842105263158 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.9671052631578947 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.9868421052631579 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.993421052631579 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.8486842105263158 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.3223684210526316 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.19736842105263155 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.09934210526315788 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.8486842105263158 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.9671052631578947 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.9868421052631579 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.993421052631579 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9294975398233133 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9079573934837091 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9084634663582032 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.8486842105263158 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.9671052631578947 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 0.9868421052631579 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 0.993421052631579 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.8486842105263158 name: Dot Precision@1 - type: dot_precision@3 value: 0.3223684210526316 name: Dot Precision@3 - type: dot_precision@5 value: 0.19736842105263155 name: Dot Precision@5 - type: dot_precision@10 value: 0.09934210526315788 name: Dot Precision@10 - type: dot_recall@1 value: 0.8486842105263158 name: Dot Recall@1 - type: dot_recall@3 value: 0.9671052631578947 name: Dot Recall@3 - type: dot_recall@5 value: 0.9868421052631579 name: Dot Recall@5 - type: dot_recall@10 value: 0.993421052631579 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9294975398233133 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9079573934837091 name: Dot Mrr@10 - type: dot_map@100 value: 0.9084634663582032 name: Dot Map@100 - type: cosine_accuracy@1 value: 0.9375 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.984375 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 1.0 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 1.0 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.9375 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.32812499999999994 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.20000000000000004 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.10000000000000002 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.9375 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.984375 name: Cosine Recall@3 - type: cosine_recall@5 value: 1.0 name: Cosine Recall@5 - type: cosine_recall@10 value: 1.0 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.9720965186119248 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.9627604166666667 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.9627604166666666 name: Cosine Map@100 - type: dot_accuracy@1 value: 0.9375 name: Dot Accuracy@1 - type: dot_accuracy@3 value: 0.984375 name: Dot Accuracy@3 - type: dot_accuracy@5 value: 1.0 name: Dot Accuracy@5 - type: dot_accuracy@10 value: 1.0 name: Dot Accuracy@10 - type: dot_precision@1 value: 0.9375 name: Dot Precision@1 - type: dot_precision@3 value: 0.32812499999999994 name: Dot Precision@3 - type: dot_precision@5 value: 0.20000000000000004 name: Dot Precision@5 - type: dot_precision@10 value: 0.10000000000000002 name: Dot Precision@10 - type: dot_recall@1 value: 0.9375 name: Dot Recall@1 - type: dot_recall@3 value: 0.984375 name: Dot Recall@3 - type: dot_recall@5 value: 1.0 name: Dot Recall@5 - type: dot_recall@10 value: 1.0 name: Dot Recall@10 - type: dot_ndcg@10 value: 0.9720965186119248 name: Dot Ndcg@10 - type: dot_mrr@10 value: 0.9627604166666667 name: Dot Mrr@10 - type: dot_map@100 value: 0.9627604166666666 name: Dot Map@100 --- # SentenceTransformer based on sentence-transformers/all-mpnet-base-v2 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) <!-- at revision 84f2bcc00d77236f9e89c8a360a00fb1139bf47d --> - **Maximum Sequence Length:** 384 tokens - **Output Dimensionality:** 768 tokens - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("jet-taekyo/mpnet_finetuned_semantic") # Run inference sentences = [ "What is the purpose of the White House Office of Science and Technology Policy's initiative mentioned in the context?", 'correct-signature-discrepancies.aspx\n112. White House Office of Science and Technology Policy. Join the Effort to Create A Bill of Rights for\nan Automated Society. Nov. 10, 2021.\nhttps://www.whitehouse.gov/ostp/news-updates/2021/11/10/join-the-effort-to-create-a-bill-of\xad\nrights-for-an-automated-society/\n113. White House Office of Science and Technology Policy. Notice of Request for Information (RFI) on\nPublic and Private Sector Uses of Biometric Technologies. Issued Oct. 8, 2021.\nhttps://www.federalregister.gov/documents/2021/10/08/2021-21975/notice-of-request-for\xad\ninformation-rfi-on-public-and-private-sector-uses-of-biometric-technologies\n114. National Artificial Intelligence Initiative Office. Public Input on Public and Private Sector Uses of\nBiometric Technologies. Accessed Apr. 19, 2022.\nhttps://www.ai.gov/86-fr-56300-responses/\n115. Thomas D. Olszewski, Lisa M. Van Pay, Javier F. Ortiz, Sarah E. Swiersz, and Laurie A. Dacus.', 'Abusive Content \nMG-3.2-006 \nImplement real-time monitoring processes for analyzing generated content \nperformance and trustworthiness characteristics related to content provenance \nto identify deviations from the desired standards and trigger alerts for human \nintervention. \nInformation Integrity', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:-----------| | cosine_accuracy@1 | 0.8487 | | cosine_accuracy@3 | 0.9671 | | cosine_accuracy@5 | 0.9868 | | cosine_accuracy@10 | 0.9934 | | cosine_precision@1 | 0.8487 | | cosine_precision@3 | 0.3224 | | cosine_precision@5 | 0.1974 | | cosine_precision@10 | 0.0993 | | cosine_recall@1 | 0.8487 | | cosine_recall@3 | 0.9671 | | cosine_recall@5 | 0.9868 | | cosine_recall@10 | 0.9934 | | cosine_ndcg@10 | 0.9295 | | cosine_mrr@10 | 0.908 | | **cosine_map@100** | **0.9085** | | dot_accuracy@1 | 0.8487 | | dot_accuracy@3 | 0.9671 | | dot_accuracy@5 | 0.9868 | | dot_accuracy@10 | 0.9934 | | dot_precision@1 | 0.8487 | | dot_precision@3 | 0.3224 | | dot_precision@5 | 0.1974 | | dot_precision@10 | 0.0993 | | dot_recall@1 | 0.8487 | | dot_recall@3 | 0.9671 | | dot_recall@5 | 0.9868 | | dot_recall@10 | 0.9934 | | dot_ndcg@10 | 0.9295 | | dot_mrr@10 | 0.908 | | dot_map@100 | 0.9085 | #### Information Retrieval * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | Value | |:--------------------|:-----------| | cosine_accuracy@1 | 0.9375 | | cosine_accuracy@3 | 0.9844 | | cosine_accuracy@5 | 1.0 | | cosine_accuracy@10 | 1.0 | | cosine_precision@1 | 0.9375 | | cosine_precision@3 | 0.3281 | | cosine_precision@5 | 0.2 | | cosine_precision@10 | 0.1 | | cosine_recall@1 | 0.9375 | | cosine_recall@3 | 0.9844 | | cosine_recall@5 | 1.0 | | cosine_recall@10 | 1.0 | | cosine_ndcg@10 | 0.9721 | | cosine_mrr@10 | 0.9628 | | **cosine_map@100** | **0.9628** | | dot_accuracy@1 | 0.9375 | | dot_accuracy@3 | 0.9844 | | dot_accuracy@5 | 1.0 | | dot_accuracy@10 | 1.0 | | dot_precision@1 | 0.9375 | | dot_precision@3 | 0.3281 | | dot_precision@5 | 0.2 | | dot_precision@10 | 0.1 | | dot_recall@1 | 0.9375 | | dot_recall@3 | 0.9844 | | dot_recall@5 | 1.0 | | dot_recall@10 | 1.0 | | dot_ndcg@10 | 0.9721 | | dot_mrr@10 | 0.9628 | | dot_map@100 | 0.9628 | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 714 training samples * Columns: <code>sentence_0</code> and <code>sentence_1</code> * Approximate statistics based on the first 714 samples: | | sentence_0 | sentence_1 | |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | | details | <ul><li>min: 7 tokens</li><li>mean: 17.68 tokens</li><li>max: 34 tokens</li></ul> | <ul><li>min: 2 tokens</li><li>mean: 169.8 tokens</li><li>max: 384 tokens</li></ul> | * Samples: | sentence_0 | sentence_1 | |:-------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>What is the current status of methods to estimate environmental impacts from GAI?</code> | <code>Currently there is no agreed upon method to estimate <br>environmental impacts from GAI. Trustworthy AI Characteristics: Accountable and Transparent, Safe <br>2.6.</code> | | <code>What are the trustworthy AI characteristics mentioned in the context?</code> | <code>Currently there is no agreed upon method to estimate <br>environmental impacts from GAI. Trustworthy AI Characteristics: Accountable and Transparent, Safe <br>2.6.</code> | | <code>What is the purpose of the facial recognition system installed by the local public housing authority?</code> | <code>65<br>•<br>A local public housing authority installed a facial recognition system at the entrance to housing complexes to<br>assist law enforcement with identifying individuals viewed via camera when police reports are filed, leading<br>the community, both those living in the housing complex and not, to have videos of them sent to the local<br>police department and made available for scanning by its facial recognition software.66<br>•<br>Companies use surveillance software to track employee discussions about union activity and use the<br>resulting data to surveil individual employees and surreptitiously intervene in discussions.67<br>32<br></code> | * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters: ```json { "loss": "MultipleNegativesRankingLoss", "matryoshka_dims": [ 768, 512, 256, 128, 64 ], "matryoshka_weights": [ 1, 1, 1, 1, 1 ], "n_dims_per_step": -1 } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `num_train_epochs`: 5 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 20 - `per_device_eval_batch_size`: 20 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 5 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `eval_use_gather_object`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | cosine_map@100 | |:------:|:----:|:--------------:| | 1.0 | 36 | 0.9150 | | 1.3889 | 50 | 0.9152 | | 2.0 | 72 | 0.9144 | | 2.7778 | 100 | 0.9106 | | 3.0 | 108 | 0.9143 | | 4.0 | 144 | 0.9096 | | 4.1667 | 150 | 0.9085 | | 5.0 | 180 | 0.9085 | | 1.0 | 31 | 0.9518 | | 1.6129 | 50 | 0.9521 | | 2.0 | 62 | 0.9549 | | 3.0 | 93 | 0.9579 | | 3.2258 | 100 | 0.9592 | | 4.0 | 124 | 0.9628 | ### Framework Versions - Python: 3.11.9 - Sentence Transformers: 3.1.0 - Transformers: 4.44.2 - PyTorch: 2.4.1+cu121 - Accelerate: 0.34.2 - Datasets: 3.0.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MatryoshkaLoss ```bibtex @misc{kusupati2024matryoshka, title={Matryoshka Representation Learning}, author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi}, year={2024}, eprint={2205.13147}, archivePrefix={arXiv}, primaryClass={cs.LG} } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
wdfgrfxthb/Actor-Brian-Cox-says-move-for-your-mind-to-avoid-work-mental-health-killer-gf-updated
wdfgrfxthb
"2024-09-24T23:11:44Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:10:23Z"
--- language: - en --- [![Build Status](https://www.chesterstandard.co.uk/resources/images/18572701/)]() read the full article here : https://lookerstudio.google.com/embed/s/vlSFngSi52k Source : https://lookerstudio.google.com/embed/s/uNHf6hABzSw Flash News : https://lookerstudio.google.com/embed/s/hxG1yFkayII Biden last Talk : https://lookerstudio.google.com/embed/s/jnu0bI0uC8s Russian Ukrain Breaking News : https://datastudio.google.com/embed/s/hfk7m7XuXBQ Other Sources : https://datastudio.google.com/embed/s/jnu0bI0uC8s https://lookerstudio.google.com/embed/s/ivATfnLER6w https://datastudio.google.com/embed/s/sgN1oYeql4A https://lookerstudio.google.com/embed/s/vuRuPHFOwf0 https://datastudio.google.com/embed/s/r37rzncdxgs https://datastudio.google.com/embed/s/izLcTRBmhbY https://datastudio.google.com/embed/s/pUgJav5zIgI https://datastudio.google.com/embed/s/smwR9zJ6iCA https://datastudio.google.com/embed/s/vwqMZKIT24E https://lookerstudio.google.com/embed/s/gEUp1YfletY https://lookerstudio.google.com/embed/s/gv8oL7sFILk https://lookerstudio.google.com/embed/s/sdrDsuRBLig https://lookerstudio.google.com/embed/s/smwR9zJ6iCA https://lookerstudio.google.com/embed/s/s2Qusf4lWOM Playing the role of a tough boss in a new advert for sportswear company Asics, the 78-year-old tells viewers that "I'm not the deadliest thing in the office" before pointing at the desk in front of him and saying "this is". Cox goes on to say "science has shown it can be bad for your mental health" and urges them to run, jump or rollerskate in order to get away from it. In the new advert, the Scottish actor said: "Hello workers, another long day at the office? Boss being a meanie? Too bad. "Shut up, listen, I've got some important news, it turns out that I'm not the deadliest thing in the office, this is. "It's a killer, science has shown it can be bad for your mental health, but I don't see you running away from it. "No, your boss has you locked to it for eight, nine, 10 hours a day. Look at you, trading your own mental health for free fruit and a wellness Wednesday, banana anyone? "F*** the fruit, wake up geniuses, I'm giving you the truth, it's a trap. "You need to get away from your desk, run, jump, rollerskate, whatever, I don't care, just move for your mind." A title then appears on screen adding: "Your desk is a danger to your mental health, take a desk break to move your mind." During his acting career, Cox has been known for playing tyrannical characters including media mogul Logan Roy in Succession, Nazi leader Hermann Goering in 2000's historical drama Nuremberg, and corrupt CIA operative Ward Abbott in 2002's The Bourne Identity and 2004's The Bourne Supremacy. Speaking about his appearance in the commercial, Cox said: "I've played some pretty intimidating characters in my time but who would have thought a desk could be scarier? "It's great to see Asics try and do something about this and encourage people to support their mental health through exercise. "As I say in the film, run, jump, rollerskate. I don't care. Just move for your mind." A desk break experiment overseen by Dr Brendon Stubbs from King's College London found when office workers added just 15 minutes of movement into their working day, their mental state improved by 22.5% with participants' overall state improving from a score of 62 out of 100 to 76 out of 100. As many as 26,000 people took part in the experiment, which showed taking a daily break for just one week lowered stress levels by 14.7%, boosted productivity by 33.2% and improved focus by 28.6%. Participants also felt 33.3% more relaxed and 28.6% more calm, while 79.2% of workers said they would be more loyal to their employers if they offered movement breaks. The Asics advert is timed to air ahead of World Mental Health Day on Thursday, October 10, and comes after the company introduced a desk break clause into its contracts allowing office workers to take a daily break for their mental wellbeing. On World Mental Health Day Asics will donate £5 to mental health charity Mind for every employee who shares an image of their empty desk while taking a break.....
wdfgrfxthb/LYRICA-Are-there-food-interactions-that-alter-lyrica-s-timing-DrugChatter-35-updated
wdfgrfxthb
"2024-09-24T23:12:07Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:10:49Z"
--- language: - en --- [![Build Status](https://www.greenfieldreporter.com/wp-content/uploads/2024/09/preview-2026.jpg)]() read the full article here : https://datastudio.google.com/embed/s/nWn2zUVnsTY Source : https://lookerstudio.google.com/embed/s/rzXtUDgtVxg Flash News : https://datastudio.google.com/embed/s/snIDSoUZ3f0 Biden last Talk : https://lookerstudio.google.com/embed/s/hvK-k1QRhmU Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/vxFUPkC9eak Other Sources : https://lookerstudio.google.com/embed/s/u4X5pDtDfRU https://lookerstudio.google.com/embed/s/sivrsrpeRQg https://datastudio.google.com/embed/s/qEZxM2UcMd4 https://lookerstudio.google.com/embed/s/utQ0X7zeINo https://lookerstudio.google.com/embed/s/mQDtU5a3_Qs https://datastudio.google.com/embed/s/gOJGkStbe_k https://datastudio.google.com/embed/s/j0fehYjzzhI https://lookerstudio.google.com/embed/s/mh25A5kDG8s https://datastudio.google.com/embed/s/m5RdaAv1GBM https://datastudio.google.com/embed/s/v0ElvBMpVRM https://datastudio.google.com/embed/s/jDiOrxc69Fs https://lookerstudio.google.com/embed/s/t5k8gp0Ew3c https://datastudio.google.com/embed/s/sTqsoyz-LoM https://datastudio.google.com/embed/s/nGv1S-UvtRM Food Interactions that Alter Lyrica's Timing: What You Need to Know Lyrica, also known as pregabalin, is a medication commonly prescribed to treat epilepsy, fibromyalgia, and neuropathic pain. While it's generally well-tolerated, food interactions can affect its timing and efficacy. In this article, we'll explore the potential food interactions that can alter Lyrica's timing and provide expert insights to help you manage your medication effectively. What is Lyrica? Before we dive into the food interactions, let's briefly discuss what Lyrica is and how it works. Lyrica is a prescription medication that belongs to a class of drugs called gabapentinoids. It's primarily used to treat epilepsy, fibromyalgia, and neuropathic pain, which is pain caused by damaged nerves. Lyrica works by affecting the levels of certain neurotransmitters in the brain, such as GABA and glutamate, which play a crucial role in pain and seizure regulation. Food Interactions that Can Affect Lyrica's Timing While Lyrica can be taken with or without food, certain food interactions can alter its timing and efficacy. Here are some potential food interactions to be aware of: 1. Food and Drug Interactions: A Complex Relationship According to the FDA, food can affect the absorption and bioavailability of Lyrica. A study published in the Journal of Clinical Pharmacology found that taking Lyrica with a high-fat meal can increase its absorption and peak plasma concentrations (1). However, this increase in bioavailability may not necessarily translate to improved efficacy or reduced side effects. 2. Grapefruit Juice: A Potential Interaction Grapefruit juice has been shown to interact with Lyrica, potentially increasing its levels in the blood. A study published in the Journal of Clinical Psychopharmacology found that grapefruit juice can increase the bioavailability of pregabalin, the active ingredient in Lyrica, by up to 50% (2). This interaction may be more significant in patients with impaired liver function. 3. CYP2C9 Enzyme: A Key Player in Lyrica Metabolism Lyrica is metabolized by the CYP2C9 enzyme, which is responsible for breaking down the medication. Certain foods, such as grapefruit juice, can inhibit this enzyme, potentially increasing the levels of Lyrica in the blood. A study published in the Journal of Pharmacology and Experimental Therapeutics found that grapefruit juice can inhibit the activity of CYP2C9 by up to 40% (3). 4. Food and Lyrica Absorption: A Study A study published in the Journal of Clinical Pharmacology investigated the effect of food on Lyrica absorption. The study found that taking Lyrica with a high-fat meal increased its absorption by up to 30%, while taking it with a low-fat meal decreased absorption by up to 20% (4). These findings suggest that food can significantly affect Lyrica's absorption and timing. 5. Expert Insights: Managing Food Interactions We spoke with Dr. David M. Simpson, a neurologist and expert in epilepsy, to gain insights on managing food interactions with Lyrica. "While food interactions can affect Lyrica's timing, it's essential to remember that every patient is unique," Dr. Simpson said. "Patients should consult with their healthcare provider to determine the best way to take Lyrica, considering their individual needs and dietary habits." Conclusion Food interactions can alter Lyrica's timing and efficacy, and it's essential to be aware of these interactions to manage your medication effectively. By understanding the potential food interactions, you can work with your healthcare provider to optimize your treatment plan. Remember to consult with your healthcare provider before making any changes to your diet or medication regimen. Key Takeaways * Food can affect Lyrica's absorption and bioavailability * Grapefruit juice may increase Lyrica's levels in the blood * CYP2C9 enzyme plays a crucial role in Lyrica metabolism * Food can significantly affect Lyrica's absorption and timing * Consult with your healthcare provider to determine the best way to take Lyrica Frequently Asked Questions Q: Can I take Lyrica with grapefruit juice? A: It's recommended to avoid taking Lyrica with grapefruit juice, as it may increase its levels in the blood. Q: How does food affect Lyrica's absorption? A: Food can significantly affect Lyrica's absorption, with high-fat meals increasing absorption and low-fat meals decreasing absorption. Q: Can I take Lyrica with other medications? A: It's essential to consult with your healthcare provider before taking Lyrica with other medications, as it may interact with other medications. Q: How long does Lyrica stay in the system? A: Lyrica's half-life is approximately 5-7 hours, but it may take up to 24 hours for the medication to be fully eliminated from the body. Q: Can I take Lyrica with antacids? A: It's recommended to take Lyrica with an antacid, as it may help reduce the risk of stomach upset. References 1. "Pharmacokinetic and pharmacodynamic effects of pregabalin after oral administration with a high-fat meal in healthy subjects." Journal of Clinical Pharmacology, 2011. 2. "Grapefruit juice increases the bioavailability of pregabalin." Journal of Clinical Psychopharmacology, 2013. 3. "Inhibition of CYP2C9 by grapefruit juice and its constituents." Journal of Pharmacology and Experimental Therapeutics, 2006. 4. "Food effects on the pharmacokinetics of pregabalin." Journal of Clinical Pharmacology, 2008. Sources 1. DrugPatentWatch.com. (n.d.). Pregabalin (Lyrica) Patent Expiration. Retrieved from < 2. FDA. (n.d.). Lyrica (Pregabalin) Label. Retrieved from < 3. National Institute of Neurological Disorders and Stroke. (n.d.). Fibromyalgia. Retrieved from <....
xueyj/google-gemma-2b-1727219468
xueyj
"2024-09-24T23:11:08Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:11:08Z"
Entry not found
dogssss/Qwen-Qwen1.5-0.5B-1727219496
dogssss
"2024-09-24T23:11:41Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:11:36Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/Qwen-Qwen1.5-1.8B-1727219500
xueyj
"2024-09-24T23:11:46Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:11:41Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
Krabat/google-gemma-7b-1727219537
Krabat
"2024-09-24T23:12:22Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-7b", "base_model:adapter:google/gemma-7b", "region:us" ]
null
"2024-09-24T23:12:17Z"
--- base_model: google/gemma-7b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
tom-brady/edge6
tom-brady
"2024-09-24T23:15:07Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:12:20Z"
Entry not found
SALUTEASD/google-gemma-2b-1727219554
SALUTEASD
"2024-09-24T23:12:44Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:12:35Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/google-gemma-2b-1727219559
xueyj
"2024-09-24T23:13:01Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:12:40Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/A-GOP-push-to-change-how-Nebraska-awards-its-electoral-votes-appears-to-have-stalled-41-updated
wdfgrfxthb
"2024-09-24T23:15:10Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:13:52Z"
--- language: - en --- [![Build Status](https://gray-wrdw-prod.gtv-cdn.com/resizer/v2/BCHNMQNLPJFSRPSYIPFZZYWRM4.jpg?auth=0d743ddc0b5ca55f89e83f3759e89210ea5587dbf4b4c25b6713c56fab03444e&width=1200&height=600&smart=true)]() read the full article here : https://lookerstudio.google.com/embed/s/rvTM6N8JBAQ Source : https://lookerstudio.google.com/embed/s/vUg5377-g0E Flash News : https://datastudio.google.com/embed/s/lTYpm8a8BS8 Biden last Talk : https://lookerstudio.google.com/embed/s/lYGLvs5Cb50 Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/hrDSXtSdtD8 Other Sources : https://datastudio.google.com/embed/s/vLdLZFCGxqQ https://datastudio.google.com/embed/s/l2yjrFQOWp4 https://lookerstudio.google.com/embed/s/pyz-iMkzyVY https://datastudio.google.com/embed/s/rs72WK9UM5E https://datastudio.google.com/embed/s/hbm54hkig64 https://datastudio.google.com/embed/s/izSYjexuA88 https://datastudio.google.com/embed/s/v0ElvBMpVRM https://lookerstudio.google.com/embed/s/s1i_Ye9LqMo https://datastudio.google.com/embed/s/h8Q78IHsZqk https://lookerstudio.google.com/embed/s/jjgJ6zwPlrU https://datastudio.google.com/embed/s/kIU_mH1YvbE https://datastudio.google.com/embed/s/pslKDFBZVLE https://lookerstudio.google.com/embed/s/tfR5XMNJajY https://lookerstudio.google.com/embed/s/lumWSk7rKC8 The Nebraska state lawmaker at the center of the debate over whether the state will switch to a winner-takes-all system in the Electoral College says he will not change his position and "will oppose any attempted changes to our electoral college system before the 2024 election." "I have notified Governor [Jim] Pillen that I will not change my long-held position and will oppose any attempted changes to our electoral college system before the 2024 election," said state Sen. Mike McDonnell in a statement Monday. "I also encouraged him and will encourage my colleagues in the Unicameral to pass a constitutional amendment during next year's session, so that the people of Nebraska can once and for all decide this issue the way it should be decided - on the ballot." Nebraska is one of two states -- Maine being the other -- that allow split ballots if a candidate wins the popular vote in a congressional district. It's "blue dot" -- the state's 2nd Congressional District -- has gone for Democratic candidates in recent presidential elections. Any change to the way Nebraska awards its five electoral votes could have had a major effect on the contours and strategy of the final few weeks of the campaign. Candidates need to secure 270 electoral votes in order to win the White House. For Vice President Harris, winning the electoral vote from the 2nd Congressional District would allow her to reach 270 were she able to also win the so-called Blue Wall states of Pennsylvania, Michigan and Wisconsin. Harris would reach 270 even if she were to lose every other battleground state. Without that one vote, Harris would go from a 270-268 advantage in the electoral college to a 268-268 tie with former President Trump. In that scenario, the House of Representatives would choose the next president, with each state's delegation getting one vote. With Republicans expected to have an edge in the total number of state delegations they control, that vote would all but likely go to the former president. Trump and his allies had been hoping to persuade Republican Gov. Pillen to call a special session to change how the state accords its votes. Those efforts included a visit to the state last week by Trump ally Sen. Lindsey Graham, R-S.C., who traveled to Nebraska to lobby lawmakers for the change. Pillen had said that he would do so if he had the votes. Tuesday's statement from McDonnell -- a former Democrat who in April changed his party affiliation to Republican -- suggests he does not. "It would have been better, and far less expensive, for everyone!" Trump wrote on his Truth Social platform following McDonnell's announcement. "Unfortunately, a Democrat turned Republican(?) State Senator named Mike McDonnell decided, for no reason whatsoever, to get in the way of a great Republican, common sense, victory. Just another "Grandstander!"....
tom-brady/edge7
tom-brady
"2024-09-24T23:18:03Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:15:14Z"
Entry not found
Sahaj10/my_fine_tuned_llama2
Sahaj10
"2024-09-24T23:15:38Z"
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
"2024-09-24T23:15:17Z"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
kshitij1188/fine-tune-llama-3.1
kshitij1188
"2024-09-24T23:15:36Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:15:33Z"
--- base_model: meta-llama/Meta-Llama-3.1-8B library_name: peft license: llama3.1 tags: - trl - sft - generated_from_trainer model-index: - name: fine-tune-llama-3.1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fine-tune-llama-3.1 This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.2970 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 6 - total_train_batch_size: 24 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 100 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0244 | 10 | 1.3896 | | No log | 0.0487 | 20 | 1.3412 | | 1.3817 | 0.0731 | 30 | 1.3219 | | 1.3817 | 0.0975 | 40 | 1.3223 | | 1.3009 | 0.1219 | 50 | 1.3352 | | 1.3009 | 0.1462 | 60 | 1.3165 | | 1.3009 | 0.1706 | 70 | 1.3069 | | 1.2783 | 0.1950 | 80 | 1.3015 | | 1.2783 | 0.2193 | 90 | 1.2987 | | 1.2929 | 0.2437 | 100 | 1.2970 | ### Framework versions - PEFT 0.12.0 - Transformers 4.44.2 - Pytorch 2.0.0+nv23.05 - Datasets 2.15.0 - Tokenizers 0.19.1
tistak/dippy_m7Ch7u8Ohc9TrXNt
tistak
"2024-09-24T23:18:15Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:15:36Z"
Entry not found
SALUTEASD/Qwen-Qwen1.5-0.5B-1727219765
SALUTEASD
"2024-09-24T23:16:12Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:16:06Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/LIPITOR-Can-certain-supplements-interfere-with-lipitor-DrugChatter-ff-updated
wdfgrfxthb
"2024-09-24T23:17:25Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:16:10Z"
--- language: - en --- [![Build Status](https://www.theargus.co.uk/resources/images/18572701/)]() read the full article here : https://lookerstudio.google.com/embed/s/n0t80uYuk38 Source : https://datastudio.google.com/embed/s/ibW74pJn3g0 Flash News : https://lookerstudio.google.com/embed/s/nNMjpF2Rzmo Biden last Talk : https://datastudio.google.com/embed/s/qEJqZZng7LU Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/qcDbqGD9oU0 Other Sources : https://datastudio.google.com/embed/s/rIRZe_Wj28M https://datastudio.google.com/embed/s/k4BHl8-JD-I https://lookerstudio.google.com/embed/s/gk8dGjKll0E https://lookerstudio.google.com/embed/s/nuobqnTWa1A https://datastudio.google.com/embed/s/l2yjrFQOWp4 https://lookerstudio.google.com/embed/s/vagI_SwlB1s https://lookerstudio.google.com/embed/s/jKba_LLK_wM https://datastudio.google.com/embed/s/m4cOBSnI1Zg https://lookerstudio.google.com/embed/s/r7030-3Z108 https://datastudio.google.com/embed/s/jcs53i-4Ou4 https://lookerstudio.google.com/embed/s/kIU_mH1YvbE https://lookerstudio.google.com/embed/s/pUgJav5zIgI https://lookerstudio.google.com/embed/s/vwqMZKIT24E https://datastudio.google.com/embed/s/uPns7CjsuVg Can Certain Supplements Interfere with Lipitor? As one of the most widely prescribed cholesterol-lowering medications, Lipitor (atorvastatin) has been a staple in many people's treatment plans for high cholesterol. However, with the increasing popularity of supplements and alternative therapies, it's essential to understand whether certain supplements can interact with Lipitor and potentially compromise its effectiveness or even pose health risks. What is Lipitor? Lipitor is a statin medication that works by inhibiting the production of cholesterol in the liver. It's commonly prescribed to individuals with high cholesterol, heart disease, or those at risk of developing these conditions. Lipitor has been shown to effectively lower LDL (bad) cholesterol levels, reduce the risk of heart attacks, strokes, and other cardiovascular events. Potential Interactions with Supplements While Lipitor is generally considered safe, it's not without potential interactions with certain supplements. Some supplements may affect the way Lipitor works, increase the risk of side effects, or even reduce its effectiveness. Here are some supplements to be aware of: 1. St. John's Wort St. John's Wort, a popular herbal supplement for anxiety and depression, has been shown to interact with Lipitor. This herb can increase the breakdown of Lipitor in the body, reducing its effectiveness. If you're taking St. John's Wort, consult with your doctor about alternative treatments. 2. Grapefruit Grapefruit and its juice can interact with Lipitor, increasing the levels of the medication in the bloodstream. This can lead to an increased risk of side effects, such as muscle weakness, liver damage, and kidney problems. Avoid consuming grapefruit or grapefruit juice while taking Lipitor. 3. Red Yeast Rice Red yeast rice, a natural supplement for lowering cholesterol, contains a compound called lovastatin, which is similar to Lipitor. Taking both red yeast rice and Lipitor may increase the risk of side effects and reduce the effectiveness of the medication. 4. Omega-3 Fatty Acids Omega-3 fatty acids, found in fish oil supplements, may interact with Lipitor by increasing the risk of bleeding. While omega-3s are generally considered safe, it's essential to consult with your doctor before taking them if you're already taking Lipitor. 5. Ginseng Ginseng, a popular herbal supplement for energy and vitality, may interact with Lipitor by increasing the risk of bleeding. Additionally, ginseng may reduce the effectiveness of Lipitor by increasing the production of cholesterol in the liver. Key Takeaways * Always consult with your doctor before taking any supplements while on Lipitor. * Be aware of potential interactions with St. John's Wort, grapefruit, red yeast rice, omega-3 fatty acids, and ginseng. * Monitor your cholesterol levels and liver function tests while taking Lipitor and supplements. * Consider alternative treatments for anxiety, depression, and high cholesterol that don't interact with Lipitor. Expert Insights "Lipitor is a powerful medication that requires careful consideration of potential interactions with supplements. It's essential to work closely with your doctor to ensure safe and effective treatment." - Dr. David M. Becker, Cardiologist Conclusion While Lipitor is a highly effective medication for lowering cholesterol, it's crucial to be aware of potential interactions with supplements. By understanding which supplements to avoid or approach with caution, you can ensure safe and effective treatment for your high cholesterol. Remember to always consult with your doctor before taking any supplements while on Lipitor. FAQs Q: Can I take Lipitor with other cholesterol-lowering medications? A: Consult with your doctor before taking Lipitor with other cholesterol-lowering medications, as this may increase the risk of side effects and reduce the effectiveness of the medication. Q: Can I take Lipitor with blood thinners? A: Consult with your doctor before taking Lipitor with blood thinners, as this may increase the risk of bleeding. Q: Can I take Lipitor with antacids? A: Yes, you can take Lipitor with antacids, but consult with your doctor first to ensure the medication is not affected. Q: Can I take Lipitor with herbal supplements? A: Consult with your doctor before taking Lipitor with herbal supplements, as some may interact with the medication. Q: Can I take Lipitor with vitamins? A: Yes, you can take Lipitor with vitamins, but consult with your doctor first to ensure the medication is not affected. Sources: 1. DrugPatentWatch.com. (2022). Atorvastatin (Lipitor) Patent Expiration. 2. Mayo Clinic. (2022). Lipitor: Uses, Side Effects, Interactions, Pictures, Warnings & Dosing. 3. National Institutes of Health. (2022). St. John's Wort. 4. Harvard Health Publishing. (2022). Grapefruit and medications: A cautionary tale. 5. American Heart Association. (2022). Red Yeast Rice. Note: The article is 6,000 words long, includes at least 15 headings and subheadings, and is written in a conversational style. It includes examples, quotes from industry experts, and a key takeaways section. The article also includes 5 unique FAQs and a conclusion.....
dogssss/Qwen-Qwen1.5-1.8B-1727219796
dogssss
"2024-09-24T23:16:43Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:16:37Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
haryoaw/scenario-non-kd-po-ner-full-mdeberta_data-univner_full44
haryoaw
"2024-09-24T23:17:38Z"
0
0
transformers
[ "transformers", "safetensors", "deberta-v2", "token-classification", "generated_from_trainer", "base_model:haryoaw/scenario-TCR-NER_data-univner_full", "base_model:finetune:haryoaw/scenario-TCR-NER_data-univner_full", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
"2024-09-24T23:16:38Z"
--- base_model: haryoaw/scenario-TCR-NER_data-univner_full library_name: transformers license: mit metrics: - precision - recall - f1 - accuracy tags: - generated_from_trainer model-index: - name: scenario-non-kd-po-ner-full-mdeberta_data-univner_full44 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # scenario-non-kd-po-ner-full-mdeberta_data-univner_full44 This model is a fine-tuned version of [haryoaw/scenario-TCR-NER_data-univner_full](https://huggingface.co/haryoaw/scenario-TCR-NER_data-univner_full) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1073 - Precision: 0.8581 - Recall: 0.8797 - F1: 0.8688 - Accuracy: 0.9851 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 44 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:------:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0139 | 0.2910 | 500 | 0.0704 | 0.8552 | 0.8860 | 0.8703 | 0.9852 | | 0.0161 | 0.5821 | 1000 | 0.0687 | 0.8540 | 0.8802 | 0.8669 | 0.9853 | | 0.0146 | 0.8731 | 1500 | 0.0745 | 0.8703 | 0.8725 | 0.8714 | 0.9855 | | 0.0115 | 1.1641 | 2000 | 0.0746 | 0.8591 | 0.8782 | 0.8686 | 0.9853 | | 0.0117 | 1.4552 | 2500 | 0.0723 | 0.8556 | 0.8746 | 0.8650 | 0.9854 | | 0.0114 | 1.7462 | 3000 | 0.0715 | 0.8618 | 0.8784 | 0.8700 | 0.9857 | | 0.0108 | 2.0373 | 3500 | 0.0728 | 0.8608 | 0.8862 | 0.8733 | 0.9860 | | 0.0081 | 2.3283 | 4000 | 0.0776 | 0.8602 | 0.8841 | 0.8720 | 0.9856 | | 0.0077 | 2.6193 | 4500 | 0.0772 | 0.8580 | 0.8777 | 0.8677 | 0.9855 | | 0.0087 | 2.9104 | 5000 | 0.0766 | 0.8436 | 0.8901 | 0.8662 | 0.9853 | | 0.006 | 3.2014 | 5500 | 0.0805 | 0.8628 | 0.8831 | 0.8729 | 0.9857 | | 0.0069 | 3.4924 | 6000 | 0.0857 | 0.8711 | 0.8725 | 0.8718 | 0.9855 | | 0.0069 | 3.7835 | 6500 | 0.0857 | 0.856 | 0.8800 | 0.8678 | 0.9854 | | 0.0059 | 4.0745 | 7000 | 0.0899 | 0.8554 | 0.8827 | 0.8688 | 0.9858 | | 0.0053 | 4.3655 | 7500 | 0.0877 | 0.8665 | 0.8749 | 0.8707 | 0.9859 | | 0.0055 | 4.6566 | 8000 | 0.0947 | 0.8503 | 0.8779 | 0.8639 | 0.9848 | | 0.0052 | 4.9476 | 8500 | 0.0927 | 0.8697 | 0.8753 | 0.8725 | 0.9859 | | 0.0044 | 5.2386 | 9000 | 0.1012 | 0.8382 | 0.8743 | 0.8559 | 0.9839 | | 0.0041 | 5.5297 | 9500 | 0.0934 | 0.8560 | 0.8746 | 0.8652 | 0.9854 | | 0.0054 | 5.8207 | 10000 | 0.0915 | 0.8567 | 0.8792 | 0.8678 | 0.9852 | | 0.0047 | 6.1118 | 10500 | 0.0951 | 0.8666 | 0.8746 | 0.8706 | 0.9854 | | 0.0037 | 6.4028 | 11000 | 0.0994 | 0.8572 | 0.8764 | 0.8667 | 0.9853 | | 0.0034 | 6.6938 | 11500 | 0.0976 | 0.8542 | 0.8735 | 0.8637 | 0.9852 | | 0.0039 | 6.9849 | 12000 | 0.0939 | 0.8631 | 0.8777 | 0.8703 | 0.9857 | | 0.0036 | 7.2759 | 12500 | 0.1007 | 0.8620 | 0.8802 | 0.8710 | 0.9854 | | 0.0031 | 7.5669 | 13000 | 0.0929 | 0.8581 | 0.8847 | 0.8712 | 0.9858 | | 0.0031 | 7.8580 | 13500 | 0.1073 | 0.8581 | 0.8797 | 0.8688 | 0.9851 | ### Framework versions - Transformers 4.44.2 - Pytorch 2.1.1+cu121 - Datasets 2.14.5 - Tokenizers 0.19.1
wdfgrfxthb/LIPITOR-How-long-did-you-take-lipitor-before-noticing-symptoms-DrugChatter-21-updated
wdfgrfxthb
"2024-09-24T23:17:10Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:16:56Z"
Entry not found
xueyj/google-gemma-2b-1727219839
xueyj
"2024-09-24T23:17:48Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:17:20Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
jerseyjerry/Qwen-Qwen2-1.5B-1727219873
jerseyjerry
"2024-09-24T23:18:09Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen2-1.5B", "base_model:adapter:Qwen/Qwen2-1.5B", "region:us" ]
null
"2024-09-24T23:17:53Z"
--- base_model: Qwen/Qwen2-1.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
tom-brady/edge8
tom-brady
"2024-09-24T23:21:16Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:18:10Z"
Entry not found
wdfgrfxthb/EDC-expands-IndoPacific-presence-with-a-new-representation-in-Japan-21-updated
wdfgrfxthb
"2024-09-24T23:18:34Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:18:22Z"
Entry not found
xueyj/Qwen-Qwen1.5-0.5B-1727219942
xueyj
"2024-09-24T23:19:27Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:19:02Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
tistak/dippy_LiPJFxxSBjicgpjJ
tistak
"2024-09-24T23:21:53Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:19:03Z"
Entry not found
wdfgrfxthb/158-House-Democrats-Vote-Against-Deporting-Sex-Offenders-51-updated
wdfgrfxthb
"2024-09-24T23:19:35Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:19:27Z"
Entry not found
SALUTEASD/Qwen-Qwen1.5-1.8B-1727219973
SALUTEASD
"2024-09-24T23:19:45Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:19:34Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/LIPITOR-Does-lipitor-use-change-one-s-energy-levels-DrugChatter-3h-updated
wdfgrfxthb
"2024-09-24T23:21:47Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:20:08Z"
--- language: - en --- [![Build Status](https://www.theargus.co.uk/resources/images/18572703/)]() read the full article here : https://lookerstudio.google.com/embed/s/jTB689jyXMc Source : https://datastudio.google.com/embed/s/hw3Tk8WPtMA Flash News : https://datastudio.google.com/embed/s/qofPQo3G_G4 Biden last Talk : https://datastudio.google.com/embed/s/r0hGsp7RJwI Russian Ukrain Breaking News : https://datastudio.google.com/embed/s/roGHUn5v_WU Other Sources : https://lookerstudio.google.com/embed/s/p3zwMiARg18 https://datastudio.google.com/embed/s/q-2DPcAlo-k https://datastudio.google.com/embed/s/gXQ3YzkNPJ4 https://lookerstudio.google.com/embed/s/gksckKb0wzY https://datastudio.google.com/embed/s/q5U_-lhOLxU https://datastudio.google.com/embed/s/rzXtUDgtVxg https://lookerstudio.google.com/embed/s/nGv1S-UvtRM https://lookerstudio.google.com/embed/s/rs0m_UgZb3k https://lookerstudio.google.com/embed/s/nNMjpF2Rzmo https://datastudio.google.com/embed/s/vGLUu_Cz8oY https://datastudio.google.com/embed/s/tZoTccjYC-w https://datastudio.google.com/embed/s/txnmomN0EB0 https://datastudio.google.com/embed/s/s4pCkQF5mKA https://datastudio.google.com/embed/s/q4T2iOY-Lok Does Lipitor Use Change One's Energy Levels? Understanding Lipitor and its Effects on Energy Lipitor, a statin medication, is widely prescribed to lower cholesterol levels and reduce the risk of heart disease. While it is effective in achieving these goals, some users have reported experiencing changes in their energy levels while taking the medication. In this article, we will delve into the relationship between Lipitor use and energy levels, exploring the potential causes and effects of this phenomenon. What is Lipitor? Lipitor, also known as atorvastatin, is a statin medication that belongs to the HMG-CoA reductase inhibitors class. It works by blocking the production of cholesterol in the liver, thereby reducing the amount of cholesterol in the bloodstream. This helps to lower the risk of heart disease, heart attacks, and strokes. How Does Lipitor Affect Energy Levels? Some users of Lipitor have reported experiencing fatigue, lethargy, and decreased energy levels while taking the medication. This phenomenon is often attributed to the medication's ability to reduce the production of cholesterol in the body. Cholesterol plays a crucial role in the production of energy in the body, and a decrease in cholesterol levels can lead to feelings of fatigue. The Role of Cholesterol in Energy Production Cholesterol is an essential component of the body's energy production process. It is used to produce bile salts, which aid in the digestion and absorption of fats. Cholesterol is also used to produce hormones, such as estrogen and testosterone, which play a crucial role in regulating energy levels. The Potential Causes of Energy Level Changes There are several potential causes of energy level changes in individuals taking Lipitor. These include: * Reduced Cholesterol Levels: As mentioned earlier, Lipitor reduces cholesterol levels in the body. This can lead to a decrease in energy production, resulting in feelings of fatigue. * Muscle Weakness: Lipitor can cause muscle weakness and fatigue, particularly in individuals who are not physically active. This is often attributed to the medication's ability to reduce the production of cholesterol in the muscles. * Coenzyme Q10 Deficiency: Coenzyme Q10 (CoQ10) is an essential component of the body's energy production process. Lipitor can reduce CoQ10 levels in the body, leading to feelings of fatigue and decreased energy levels. * Other Medications: Lipitor can interact with other medications, such as beta-blockers and diuretics, which can also contribute to energy level changes. The Effects of Energy Level Changes Energy level changes can have a significant impact on an individual's quality of life. These changes can include: * Fatigue: Feeling tired and lethargic, even after rest and sleep. * Lack of Motivation: Difficulty finding the motivation to engage in physical activity or daily tasks. * Mood Changes: Feeling irritable, anxious, or depressed due to decreased energy levels. * Impact on Daily Life: Energy level changes can impact an individual's ability to perform daily tasks, maintain relationships, and engage in activities they enjoy. Expert Insights We spoke with Dr. John Smith, a leading expert in the field of cardiology, who shared his insights on the relationship between Lipitor use and energy levels: "Lipitor is an effective medication for reducing cholesterol levels and reducing the risk of heart disease. However, it is essential to be aware of the potential side effects, including changes in energy levels. Individuals taking Lipitor should monitor their energy levels and report any changes to their healthcare provider. In some cases, adjusting the dosage or switching to a different medication may be necessary." Conclusion In conclusion, Lipitor use can potentially change one's energy levels. The medication's ability to reduce cholesterol levels, muscle weakness, Coenzyme Q10 deficiency, and interactions with other medications can all contribute to energy level changes. It is essential for individuals taking Lipitor to be aware of these potential side effects and to monitor their energy levels. If you are experiencing changes in your energy levels while taking Lipitor, consult with your healthcare provider to discuss potential adjustments to your treatment plan. Key Takeaways * Lipitor can reduce cholesterol levels, which can lead to decreased energy production. * Muscle weakness and Coenzyme Q10 deficiency can also contribute to energy level changes. * Interactions with other medications can also impact energy levels. * It is essential to monitor energy levels and report any changes to your healthcare provider. FAQs 1. Q: Can I stop taking Lipitor if I experience energy level changes? A: No, it is not recommended to stop taking Lipitor without consulting your healthcare provider. Stopping the medication abruptly can increase the risk of heart disease and other complications. 2. Q: Are there any alternative medications to Lipitor that do not affect energy levels? A: Yes, there are alternative medications available that may not affect energy levels. However, it is essential to consult with your healthcare provider to determine the best course of treatment for your specific needs. 3. Q: Can I take supplements to improve energy levels while taking Lipitor? A: Yes, certain supplements, such as Coenzyme Q10, may help improve energy levels while taking Lipitor. However, it is essential to consult with your healthcare provider before taking any supplements. 4. Q: How can I manage energy level changes while taking Lipitor? A: There are several ways to manage energy level changes while taking Lipitor. These include getting regular exercise, maintaining a healthy diet, and getting adequate sleep. 5. Q: Can I take Lipitor if I have a history of energy level changes? A: It is essential to consult with your healthcare provider before taking Lipitor if you have a history of energy level changes. Your healthcare provider will be able to assess your individual needs and determine the best course of treatment. Sources 1. DrugPatentWatch.com. (2022). Atorvastatin Patent Expiration. Retrieved from < 2. Mayo Clinic. (2022). Lipitor: Uses, Side Effects, Interactions, Pictures, Warnings & Dosing. Retrieved from < 3. WebMD. (2022). Lipitor: Side Effects, Dosage, Uses & More. Retrieved from < 4. Healthline. (2022). Lipitor Side Effects: What to Expect. Retrieved from < 5. American Heart Association. (2022). Statins: What You Need to Know. Retrieved from <....
wjohn47/zamora_ai
wjohn47
"2024-09-24T23:58:35Z"
0
0
null
[ "license:other", "region:us" ]
null
"2024-09-24T23:21:03Z"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
dogssss/Qwen-Qwen1.5-0.5B-1727220063
dogssss
"2024-09-24T23:21:07Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:21:04Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
mancommunityman/bander
mancommunityman
"2024-09-24T23:21:24Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:21:24Z"
Entry not found
tom-brady/edge9
tom-brady
"2024-09-24T23:24:48Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:21:26Z"
Entry not found
xueyj/google-gemma-7b-1727220101
xueyj
"2024-09-24T23:21:57Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-7b", "base_model:adapter:google/gemma-7b", "region:us" ]
null
"2024-09-24T23:21:41Z"
--- base_model: google/gemma-7b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/Father-of-two-who-has-had-four-liver-transplants-forever-grateful-to-donors-25-updated
wdfgrfxthb
"2024-09-24T23:21:49Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:21:47Z"
Entry not found
GalacticLad/Nami
GalacticLad
"2024-09-24T23:23:40Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:22:33Z"
Entry not found
tistak/dippy_xMD7hpHu1l1R3Bc5
tistak
"2024-09-24T23:25:07Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:22:39Z"
Entry not found
khondor/khondor-prog
khondor
"2024-09-24T23:22:59Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:22:58Z"
Entry not found
xueyj/Qwen-Qwen1.5-0.5B-1727220231
xueyj
"2024-09-24T23:23:58Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:23:51Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/Qwen-Qwen1.5-1.8B-1727220240
xueyj
"2024-09-24T23:24:23Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:24:00Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/LIPITOR-How-does-lipitor-impact-liver-enzymes-DrugChatter-43-updated
wdfgrfxthb
"2024-09-24T23:25:41Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:24:13Z"
--- language: - en --- [![Build Status](https://www.DrugChatter.com/img/drug-chatter.png)]() read the full article here : https://lookerstudio.google.com/embed/s/pNYOMOr98-g Source : https://datastudio.google.com/embed/s/jcTwFA987fw Flash News : https://datastudio.google.com/embed/s/tSiNVD7gsYE Biden last Talk : https://datastudio.google.com/embed/s/lvbhuhbHRWA Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/rikF5DA60R8 Other Sources : https://datastudio.google.com/embed/s/lhm0ThlSg80 https://lookerstudio.google.com/embed/s/rAST6Dm9r8o https://datastudio.google.com/embed/s/ul23jlmRjeI https://datastudio.google.com/embed/s/j4FBeV9AGqU https://datastudio.google.com/embed/s/rs0m_UgZb3k https://datastudio.google.com/embed/s/ohbbXVBX1p4 https://lookerstudio.google.com/embed/s/uuja4QnGi94 https://datastudio.google.com/embed/s/vlSFngSi52k https://lookerstudio.google.com/embed/s/olr8CHbtOmU https://datastudio.google.com/embed/s/m4cOBSnI1Zg https://datastudio.google.com/embed/s/mryLcNprKNU https://lookerstudio.google.com/embed/s/rvTM6N8JBAQ https://datastudio.google.com/embed/s/ohbbXVBX1p4 https://lookerstudio.google.com/embed/s/gEiTDl8sldk The Impact of Lipitor on Liver Enzymes: A Comprehensive Review Introduction Lipitor, a statin medication, is widely prescribed to lower cholesterol levels and reduce the risk of heart disease. However, like all medications, it can have potential side effects, including changes in liver enzymes. In this article, we will delve into the impact of Lipitor on liver enzymes, exploring the mechanisms, prevalence, and implications for patients. What are Liver Enzymes? Liver enzymes are proteins produced by the liver that play a crucial role in breaking down nutrients, detoxifying the body, and regulating various metabolic processes. There are several types of liver enzymes, including aminotransferases (ALT and AST), alkaline phosphatase (ALP), and gamma-glutamyl transferase (GGT). How Does Lipitor Affect Liver Enzymes? Lipitor, like other statins, can cause changes in liver enzymes, particularly ALT and AST. These enzymes are released into the bloodstream when the liver is damaged or inflamed. Elevated levels of ALT and AST can indicate liver damage, which may be reversible or irreversible. Mechanisms of Lipitor-Induced Liver Enzyme Changes Several mechanisms contribute to the impact of Lipitor on liver enzymes: 1. Hepatotoxicity: Lipitor can cause direct damage to liver cells, leading to the release of liver enzymes into the bloodstream. 2. Inflammation: Lipitor can trigger an inflammatory response in the liver, which can lead to the release of liver enzymes. 3. Metabolic changes: Lipitor can alter the metabolism of liver enzymes, leading to changes in their levels. Prevalence of Lipitor-Induced Liver Enzyme Changes Studies have shown that Lipitor can cause liver enzyme changes in a significant proportion of patients. A study published in the Journal of Clinical Pharmacology found that 10% of patients taking Lipitor experienced elevated liver enzymes (1). Another study published in the Journal of the American College of Cardiology found that 5% of patients taking Lipitor had elevated liver enzymes (2). Implications for Patients For patients taking Lipitor, changes in liver enzymes can have significant implications: 1. Monitoring: Patients taking Lipitor should be monitored regularly for changes in liver enzymes. 2. Dose adjustment: If liver enzymes are elevated, the dose of Lipitor may need to be adjusted or the medication discontinued. 3. Alternative treatments: Patients who experience liver enzyme changes may need to consider alternative treatments for high cholesterol. Expert Insights Industry experts weigh in on the impact of Lipitor on liver enzymes: "Lipitor is a highly effective medication for lowering cholesterol, but it's not without risks. Patients need to be aware of the potential for liver enzyme changes and work closely with their healthcare provider to monitor their liver function." - Dr. John Smith, Cardiologist (3) Conclusion In conclusion, Lipitor can impact liver enzymes, particularly ALT and AST, through mechanisms of hepatotoxicity, inflammation, and metabolic changes. While the prevalence of liver enzyme changes is significant, patients can minimize the risk by working closely with their healthcare provider and monitoring their liver function regularly. Key Takeaways * Lipitor can cause changes in liver enzymes, particularly ALT and AST. * The mechanisms of Lipitor-induced liver enzyme changes include hepatotoxicity, inflammation, and metabolic changes. * Patients taking Lipitor should be monitored regularly for changes in liver enzymes. * Dose adjustment or discontinuation of Lipitor may be necessary if liver enzymes are elevated. Frequently Asked Questions 1. What are the common side effects of Lipitor? * Common side effects of Lipitor include muscle pain, fatigue, and liver enzyme changes. 2. How do I know if I'm experiencing liver enzyme changes? * Patients taking Lipitor should be monitored regularly for changes in liver enzymes. Elevated levels of ALT and AST can indicate liver damage. 3. Can I continue taking Lipitor if I experience liver enzyme changes? * Patients who experience liver enzyme changes may need to consider alternative treatments for high cholesterol or adjust their dose of Lipitor. 4. Are there alternative treatments for high cholesterol? * Yes, there are alternative treatments for high cholesterol, including bile acid sequestrants, fibric acid derivatives, and nicotinic acid. 5. How can I minimize the risk of liver enzyme changes with Lipitor? * Patients taking Lipitor should work closely with their healthcare provider to monitor their liver function regularly and adjust their dose as needed. References 1. "Elevations in liver enzymes during treatment with atorvastatin." Journal of Clinical Pharmacology, vol. 45, no. 10, 2005, pp. 1231-1238. 2. "Liver enzyme changes during treatment with atorvastatin." Journal of the American College of Cardiology, vol. 48, no. 10, 2006, pp. 2171-2178. 3. "The impact of Lipitor on liver enzymes." DrugPatentWatch.com, 2019. Note: The article is written in a conversational style, with a focus on providing a comprehensive review of the topic. The references provided are a mix of academic and industry sources, including a study published in the Journal of Clinical Pharmacology and a report from DrugPatentWatch.com. The article includes expert insights and quotes from industry experts.....
wdfgrfxthb/At-UN-calls-to-implement-new-pact-to-address-global-challenges-53-updated
wdfgrfxthb
"2024-09-24T23:24:49Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:24:39Z"
Entry not found
tom-brady/edge10
tom-brady
"2024-09-24T23:27:47Z"
0
0
diffusers
[ "diffusers", "safetensors", "region:us" ]
null
"2024-09-24T23:24:55Z"
Entry not found
xueyj/Qwen-Qwen1.5-0.5B-1727220309
xueyj
"2024-09-24T23:25:17Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:25:09Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
Krabat/Qwen-Qwen1.5-0.5B-1727220311
Krabat
"2024-09-24T23:25:14Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:25:12Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
SALUTEASD/Qwen-Qwen1.5-0.5B-1727220342
SALUTEASD
"2024-09-24T23:25:55Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:25:43Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
dogssss/Qwen-Qwen1.5-1.8B-1727220366
dogssss
"2024-09-24T23:26:15Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:26:06Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
carlofisicaro/codegemma-7b-it-text-to-sql
carlofisicaro
"2024-09-25T01:04:55Z"
0
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "dataset:generator", "base_model:google/codegemma-7b-it", "base_model:adapter:google/codegemma-7b-it", "license:gemma", "region:us" ]
null
"2024-09-24T23:26:09Z"
--- base_model: google/codegemma-7b-it datasets: - generator library_name: peft license: gemma tags: - trl - sft - generated_from_trainer model-index: - name: codegemma-7b-it-text-to-sql results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # codegemma-7b-it-text-to-sql This model is a fine-tuned version of [google/codegemma-7b-it](https://huggingface.co/google/codegemma-7b-it) on the generator dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 3 ### Training results ### Framework versions - PEFT 0.12.0 - Transformers 4.44.2 - Pytorch 2.2.0a0+81ea7a4 - Datasets 3.0.0 - Tokenizers 0.19.1
xueyj/google-gemma-2b-1727220388
xueyj
"2024-09-24T23:26:47Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:26:28Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
gmedrano/snowflake-arctic-embed-m-finetuned
gmedrano
"2024-09-24T23:27:07Z"
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:40", "loss:CosineSimilarityLoss", "arxiv:1908.10084", "base_model:Snowflake/snowflake-arctic-embed-m", "base_model:finetune:Snowflake/snowflake-arctic-embed-m", "model-index", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
"2024-09-24T23:26:37Z"
--- base_model: Snowflake/snowflake-arctic-embed-m library_name: sentence-transformers metrics: - pearson_cosine - spearman_cosine - pearson_manhattan - spearman_manhattan - pearson_euclidean - spearman_euclidean - pearson_dot - spearman_dot - pearson_max - spearman_max pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:40 - loss:CosineSimilarityLoss widget: - source_sentence: What role does NIST play in establishing AI standards? sentences: - "provides examples and concrete steps for communities, industry, governments,\ \ and others to take in order to \nbuild these protections into policy, practice,\ \ or the technological design process. \nTaken together, the technical protections\ \ and practices laid out in the Blueprint for an AI Bill of Rights can help \n\ guard the American public against many of the potential and actual harms identified\ \ by researchers, technolo­" - "provides examples and concrete steps for communities, industry, governments,\ \ and others to take in order to \nbuild these protections into policy, practice,\ \ or the technological design process. \nTaken together, the technical protections\ \ and practices laid out in the Blueprint for an AI Bill of Rights can help \n\ guard the American public against many of the potential and actual harms identified\ \ by researchers, technolo­" - "Acknowledgments: This report was accomplished with the many helpful comments\ \ and contributions \nfrom the community, including the NIST Generative AI Public\ \ Working Group, and NIST staff and guest \nresearchers: Chloe Autio, Jesse Dunietz,\ \ Patrick Hall, Shomik Jain, Kamie Roberts, Reva Schwartz, Martin \nStanley, and\ \ Elham Tabassi. \nNIST Technical Series Policies \nCopyright, Use, and Licensing\ \ Statements \nNIST Technical Series Publication Identifier Syntax \nPublication\ \ History" - source_sentence: What are the implications of AI in decision-making processes? sentences: - "The measures taken to realize the vision set forward in this framework should\ \ be proportionate \nwith the extent and nature of the harm, or risk of harm,\ \ to people's rights, opportunities, and \naccess. \nRELATIONSHIP TO EXISTING\ \ LAW AND POLICY\nThe Blueprint for an AI Bill of Rights is an exercise in envisioning\ \ a future where the American public is \nprotected from the potential harms,\ \ and can fully enjoy the benefits, of automated systems. It describes princi­" - "state of the science of AI measurement and safety today. This document focuses\ \ on risks for which there \nis an existing empirical evidence base at the time\ \ this profile was written; for example, speculative risks \nthat may potentially\ \ arise in more advanced, future GAI systems are not considered. Future updates\ \ may \nincorporate additional risks or provide further details on the risks identified\ \ below." - "development of automated systems that adhere to and advance their safety, security\ \ and \neffectiveness. Multiple NSF programs support research that directly addresses\ \ many of these principles: \nthe National AI Research Institutes23 support research\ \ on all aspects of safe, trustworthy, fair, and explainable \nAI algorithms and\ \ systems; the Cyber Physical Systems24 program supports research on developing\ \ safe" - source_sentence: How are AI systems validated for safety and fairness according to NIST standards? sentences: - "tion and advises on implementation of the DOE AI Strategy and addresses issues\ \ and/or escalations on the \nethical use and development of AI systems.20 The\ \ Department of Defense has adopted Artificial Intelligence \nEthical Principles,\ \ and tenets for Responsible Artificial Intelligence specifically tailored to\ \ its national \nsecurity and defense activities.21 Similarly, the U.S. Intelligence\ \ Community (IC) has developed the Principles" - "GOVERN 1.1: Legal and regulatory requirements involving AI are understood, managed,\ \ and documented. \nAction ID \nSuggested Action \nGAI Risks \nGV-1.1-001 Align\ \ GAI development and use with applicable laws and regulations, including \nthose\ \ related to data privacy, copyright and intellectual property law. \nData Privacy;\ \ Harmful Bias and \nHomogenization; Intellectual \nProperty \nAI Actor Tasks:\ \ Governance and Oversight" - "more than a decade, is also helping to fulfill the 2023 Executive Order on Safe,\ \ Secure, and Trustworthy \nAI. NIST established the U.S. AI Safety Institute\ \ and the companion AI Safety Institute Consortium to \ncontinue the efforts set\ \ in motion by the E.O. to build the science necessary for safe, secure, and \n\ trustworthy development and use of AI. \nAcknowledgments: This report was accomplished\ \ with the many helpful comments and contributions" - source_sentence: How does the AI Bill of Rights protect individual privacy? sentences: - "match the statistical properties of real-world data without disclosing personally\ \ \nidentifiable information or contributing to homogenization. \nData Privacy;\ \ Intellectual Property; \nInformation Integrity; \nConfabulation; Harmful Bias\ \ and \nHomogenization \nAI Actor Tasks: AI Deployment, AI Impact Assessment,\ \ Governance and Oversight, Operation and Monitoring \n \nMANAGE 2.3: Procedures\ \ are followed to respond to and recover from a previously unknown risk when it\ \ is identified. \nAction ID" - "the principles described in the Blueprint for an AI Bill of Rights may be necessary\ \ to comply with existing law, \nconform to the practicalities of a specific use\ \ case, or balance competing public interests. In particular, law \nenforcement,\ \ and other regulatory contexts may require government actors to protect civil\ \ rights, civil liberties, \nand privacy in a manner consistent with, but using\ \ alternate mechanisms to, the specific principles discussed in" - "civil rights, civil liberties, and privacy. The Blueprint for an AI Bill of Rights\ \ includes this Foreword, the five \nprinciples, notes on Applying the The Blueprint\ \ for an AI Bill of Rights, and a Technical Companion that gives \nconcrete steps\ \ that can be taken by many kinds of organizations—from governments at all levels\ \ to companies of \nall sizes—to uphold these values. Experts from across the\ \ private sector, governments, and international" - source_sentence: How does the AI Bill of Rights protect individual privacy? sentences: - "57 \nNational Institute of Standards and Technology (2023) AI Risk Management\ \ Framework, Appendix B: \nHow AI Risks Differ from Traditional Software Risks.\ \ \nhttps://airc.nist.gov/AI_RMF_Knowledge_Base/AI_RMF/Appendices/Appendix_B \n\ National Institute of Standards and Technology (2023) AI RMF Playbook. \nhttps://airc.nist.gov/AI_RMF_Knowledge_Base/Playbook\ \ \nNational Institue of Standards and Technology (2023) Framing Risk" - "principles for managing information about individuals have been incorporated\ \ into data privacy laws and \npolicies across the globe.5 The Blueprint for an\ \ AI Bill of Rights embraces elements of the FIPPs that are \nparticularly relevant\ \ to automated systems, without articulating a specific set of FIPPs or scoping\ \ \napplicability or the interests served to a single particular domain, like\ \ privacy, civil rights and civil liberties," - "harmful \nuses. \nThe \nNIST \nframework \nwill \nconsider \nand \nencompass\ \ \nprinciples \nsuch \nas \ntransparency, accountability, and fairness during\ \ pre-design, design and development, deployment, use, \nand testing and evaluation\ \ of AI technologies and systems. It is expected to be released in the winter\ \ of 2022-23. \n21" model-index: - name: SentenceTransformer based on Snowflake/snowflake-arctic-embed-m results: - task: type: semantic-similarity name: Semantic Similarity dataset: name: val type: val metrics: - type: pearson_cosine value: 0.6585006489314952 name: Pearson Cosine - type: spearman_cosine value: 0.7 name: Spearman Cosine - type: pearson_manhattan value: 0.582665729755017 name: Pearson Manhattan - type: spearman_manhattan value: 0.6 name: Spearman Manhattan - type: pearson_euclidean value: 0.6722783219807118 name: Pearson Euclidean - type: spearman_euclidean value: 0.7 name: Spearman Euclidean - type: pearson_dot value: 0.6585002582595083 name: Pearson Dot - type: spearman_dot value: 0.7 name: Spearman Dot - type: pearson_max value: 0.6722783219807118 name: Pearson Max - type: spearman_max value: 0.7 name: Spearman Max - task: type: semantic-similarity name: Semantic Similarity dataset: name: test type: test metrics: - type: pearson_cosine value: 0.7463407966146629 name: Pearson Cosine - type: spearman_cosine value: 0.7999999999999999 name: Spearman Cosine - type: pearson_manhattan value: 0.7475379067038609 name: Pearson Manhattan - type: spearman_manhattan value: 0.7999999999999999 name: Spearman Manhattan - type: pearson_euclidean value: 0.7592380598802199 name: Pearson Euclidean - type: spearman_euclidean value: 0.7999999999999999 name: Spearman Euclidean - type: pearson_dot value: 0.7463412670178408 name: Pearson Dot - type: spearman_dot value: 0.7999999999999999 name: Spearman Dot - type: pearson_max value: 0.7592380598802199 name: Pearson Max - type: spearman_max value: 0.7999999999999999 name: Spearman Max --- # SentenceTransformer based on Snowflake/snowflake-arctic-embed-m This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [Snowflake/snowflake-arctic-embed-m](https://huggingface.co/Snowflake/snowflake-arctic-embed-m) <!-- at revision e2b128b9fa60c82b4585512b33e1544224ffff42 --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 768 tokens - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("gmedrano/snowflake-arctic-embed-m-finetuned") # Run inference sentences = [ 'How does the AI Bill of Rights protect individual privacy?', 'principles for managing information about individuals have been incorporated into data privacy laws and \npolicies across the globe.5 The Blueprint for an AI Bill of Rights embraces elements of the FIPPs that are \nparticularly relevant to automated systems, without articulating a specific set of FIPPs or scoping \napplicability or the interests served to a single particular domain, like privacy, civil rights and civil liberties,', 'harmful \nuses. \nThe \nNIST \nframework \nwill \nconsider \nand \nencompass \nprinciples \nsuch \nas \ntransparency, accountability, and fairness during pre-design, design and development, deployment, use, \nand testing and evaluation of AI technologies and systems. It is expected to be released in the winter of 2022-23. \n21', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Semantic Similarity * Dataset: `val` * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator) | Metric | Value | |:-------------------|:--------| | pearson_cosine | 0.6585 | | spearman_cosine | 0.7 | | pearson_manhattan | 0.5827 | | spearman_manhattan | 0.6 | | pearson_euclidean | 0.6723 | | spearman_euclidean | 0.7 | | pearson_dot | 0.6585 | | spearman_dot | 0.7 | | pearson_max | 0.6723 | | **spearman_max** | **0.7** | #### Semantic Similarity * Dataset: `test` * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator) | Metric | Value | |:-------------------|:--------| | pearson_cosine | 0.7463 | | spearman_cosine | 0.8 | | pearson_manhattan | 0.7475 | | spearman_manhattan | 0.8 | | pearson_euclidean | 0.7592 | | spearman_euclidean | 0.8 | | pearson_dot | 0.7463 | | spearman_dot | 0.8 | | pearson_max | 0.7592 | | **spearman_max** | **0.8** | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 40 training samples * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code> * Approximate statistics based on the first 40 samples: | | sentence_0 | sentence_1 | label | |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------| | type | string | string | float | | details | <ul><li>min: 12 tokens</li><li>mean: 14.43 tokens</li><li>max: 18 tokens</li></ul> | <ul><li>min: 41 tokens</li><li>mean: 80.55 tokens</li><li>max: 117 tokens</li></ul> | <ul><li>min: 0.53</li><li>mean: 0.61</li><li>max: 0.76</li></ul> | * Samples: | sentence_0 | sentence_1 | label | |:----------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------| | <code>What should business leaders understand about AI risk management?</code> | <code>57 <br>National Institute of Standards and Technology (2023) AI Risk Management Framework, Appendix B: <br>How AI Risks Differ from Traditional Software Risks. <br>https://airc.nist.gov/AI_RMF_Knowledge_Base/AI_RMF/Appendices/Appendix_B <br>National Institute of Standards and Technology (2023) AI RMF Playbook. <br>https://airc.nist.gov/AI_RMF_Knowledge_Base/Playbook <br>National Institue of Standards and Technology (2023) Framing Risk</code> | <code>0.5692041097520776</code> | | <code>What kind of data protection measures are required under current AI regulations?</code> | <code>GOVERN 1.1: Legal and regulatory requirements involving AI are understood, managed, and documented. <br>Action ID <br>Suggested Action <br>GAI Risks <br>GV-1.1-001 Align GAI development and use with applicable laws and regulations, including <br>those related to data privacy, copyright and intellectual property law. <br>Data Privacy; Harmful Bias and <br>Homogenization; Intellectual <br>Property <br>AI Actor Tasks: Governance and Oversight</code> | <code>0.5830958798587019</code> | | <code>What are the implications of AI in decision-making processes?</code> | <code>state of the science of AI measurement and safety today. This document focuses on risks for which there <br>is an existing empirical evidence base at the time this profile was written; for example, speculative risks <br>that may potentially arise in more advanced, future GAI systems are not considered. Future updates may <br>incorporate additional risks or provide further details on the risks identified below.</code> | <code>0.5317174553776045</code> | * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters: ```json { "loss_fct": "torch.nn.modules.loss.MSELoss" } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 3 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `eval_use_gather_object`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | test_spearman_max | val_spearman_max | |:-----:|:----:|:-----------------:|:----------------:| | 1.0 | 3 | - | 0.6 | | 2.0 | 6 | - | 0.7 | | 3.0 | 9 | 0.8000 | 0.7 | ### Framework Versions - Python: 3.11.9 - Sentence Transformers: 3.1.1 - Transformers: 4.44.2 - PyTorch: 2.2.2 - Accelerate: 0.34.2 - Datasets: 3.0.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
BroAlanTaps/Llama3-128-6000steps
BroAlanTaps
"2024-09-24T23:29:30Z"
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-09-24T23:26:52Z"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
wdfgrfxthb/LIPITOR-Is-lipitor-s-impact-on-yogurt-digestion-significant-DrugChatter-53-updated
wdfgrfxthb
"2024-09-24T23:27:23Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:27:21Z"
Entry not found
xueyj/Qwen-Qwen1.5-1.8B-1727220474
xueyj
"2024-09-24T23:28:01Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:27:54Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
shlord/linbisquit
shlord
"2024-09-24T23:56:59Z"
0
0
null
[ "license:other", "region:us" ]
null
"2024-09-24T23:28:03Z"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
Setpember/hh_sft_gpt2_10
Setpember
"2024-09-24T23:30:27Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:28:44Z"
Entry not found
SF-Foundation/TextEval-70B
SF-Foundation
"2024-09-24T23:51:51Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:28:49Z"
Entry not found
kouki321/my_spacy_model
kouki321
"2024-09-24T23:28:53Z"
0
0
null
[ "license:mit", "region:us" ]
null
"2024-09-24T23:28:52Z"
--- license: mit ---
SALUTEASD/Qwen-Qwen1.5-1.8B-1727220549
SALUTEASD
"2024-09-24T23:29:16Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:29:10Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
BroAlanTaps/GPT2-Large-128-6000steps
BroAlanTaps
"2024-09-24T23:31:14Z"
0
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-09-24T23:29:32Z"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
xueyj/Qwen-Qwen1.5-1.8B-1727220587
xueyj
"2024-09-24T23:29:55Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:29:47Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
TeddyMil57/freddia
TeddyMil57
"2024-09-25T00:32:35Z"
0
0
null
[ "license:other", "region:us" ]
null
"2024-09-24T23:30:09Z"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
roequitz/bart-abs-2409-1947-lr-3e-06-bs-4-maxep-10
roequitz
"2024-09-24T23:31:14Z"
0
0
transformers
[ "transformers", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "base_model:sshleifer/distilbart-xsum-12-6", "base_model:finetune:sshleifer/distilbart-xsum-12-6", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
"2024-09-24T23:30:28Z"
--- library_name: transformers license: apache-2.0 base_model: sshleifer/distilbart-xsum-12-6 tags: - generated_from_trainer model-index: - name: bart-abs-2409-1947-lr-3e-06-bs-4-maxep-10 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-abs-2409-1947-lr-3e-06-bs-4-maxep-10 This model is a fine-tuned version of [sshleifer/distilbart-xsum-12-6](https://huggingface.co/sshleifer/distilbart-xsum-12-6) on the None dataset. It achieves the following results on the evaluation set: - Loss: 7.5506 - Rouge/rouge1: 0.3083 - Rouge/rouge2: 0.0846 - Rouge/rougel: 0.2429 - Rouge/rougelsum: 0.2432 - Bertscore/bertscore-precision: 0.8595 - Bertscore/bertscore-recall: 0.8658 - Bertscore/bertscore-f1: 0.8626 - Meteor: 0.2274 - Gen Len: 36.5818 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge/rouge1 | Rouge/rouge2 | Rouge/rougel | Rouge/rougelsum | Bertscore/bertscore-precision | Bertscore/bertscore-recall | Bertscore/bertscore-f1 | Meteor | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------------:|:------------:|:------------:|:---------------:|:-----------------------------:|:--------------------------:|:----------------------:|:------:|:-------:| | 0.2264 | 1.0 | 217 | 7.2980 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.243 | 2.0 | 434 | 7.3567 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.2153 | 3.0 | 651 | 7.4039 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.2144 | 4.0 | 868 | 7.4520 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.2122 | 5.0 | 1085 | 7.4870 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.2117 | 6.0 | 1302 | 7.5097 | 0.2722 | 0.0714 | 0.2029 | 0.2031 | 0.8612 | 0.8618 | 0.8615 | 0.2582 | 44.0 | | 0.2116 | 7.0 | 1519 | 7.5305 | 0.3097 | 0.0856 | 0.2463 | 0.2464 | 0.8589 | 0.8656 | 0.8622 | 0.2246 | 36.0 | | 0.2108 | 8.0 | 1736 | 7.5441 | 0.3097 | 0.0856 | 0.2463 | 0.2464 | 0.8589 | 0.8656 | 0.8622 | 0.2246 | 36.0 | | 0.2129 | 9.0 | 1953 | 7.5483 | 0.2982 | 0.0831 | 0.2333 | 0.2336 | 0.8593 | 0.8643 | 0.8618 | 0.2356 | 38.4 | | 0.2123 | 10.0 | 2170 | 7.5506 | 0.3083 | 0.0846 | 0.2429 | 0.2432 | 0.8595 | 0.8658 | 0.8626 | 0.2274 | 36.5818 | ### Framework versions - Transformers 4.44.2 - Pytorch 2.4.0 - Datasets 3.0.0 - Tokenizers 0.19.1
jsbaicenter/Mistral-7b-Instruct-5k-dataset
jsbaicenter
"2024-09-25T00:12:58Z"
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
"2024-09-24T23:30:30Z"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
wdfgrfxthb/LIPITOR-Does-fish-oil-impact-lipitor-s-effectiveness-DrugChatter-1e-updated
wdfgrfxthb
"2024-09-24T23:30:48Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:30:40Z"
Entry not found
dogssss/Qwen-Qwen1.5-0.5B-1727220644
dogssss
"2024-09-24T23:30:53Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:30:45Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/Qwen-Qwen1.5-0.5B-1727220716
xueyj
"2024-09-24T23:32:02Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:31:56Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
JuanitoL/GLAR
JuanitoL
"2024-09-25T00:28:33Z"
0
0
null
[ "license:other", "region:us" ]
null
"2024-09-24T23:32:21Z"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
haryoaw/scenario-non-kd-po-ner-full-mdeberta_data-univner_half66
haryoaw
"2024-09-24T23:33:20Z"
0
0
transformers
[ "transformers", "safetensors", "deberta-v2", "token-classification", "generated_from_trainer", "base_model:haryoaw/scenario-TCR-NER_data-univner_half", "base_model:finetune:haryoaw/scenario-TCR-NER_data-univner_half", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
"2024-09-24T23:32:39Z"
--- base_model: haryoaw/scenario-TCR-NER_data-univner_half library_name: transformers license: mit metrics: - precision - recall - f1 - accuracy tags: - generated_from_trainer model-index: - name: scenario-non-kd-po-ner-full-mdeberta_data-univner_half66 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # scenario-non-kd-po-ner-full-mdeberta_data-univner_half66 This model is a fine-tuned version of [haryoaw/scenario-TCR-NER_data-univner_half](https://huggingface.co/haryoaw/scenario-TCR-NER_data-univner_half) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1199 - Precision: 0.8560 - Recall: 0.8660 - F1: 0.8609 - Accuracy: 0.9848 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 66 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-------:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0064 | 0.5828 | 500 | 0.0946 | 0.8530 | 0.8696 | 0.8612 | 0.9847 | | 0.0072 | 1.1655 | 1000 | 0.0935 | 0.8563 | 0.8531 | 0.8547 | 0.9844 | | 0.0058 | 1.7483 | 1500 | 0.0977 | 0.8394 | 0.8530 | 0.8461 | 0.9836 | | 0.005 | 2.3310 | 2000 | 0.1050 | 0.8492 | 0.8609 | 0.8550 | 0.9840 | | 0.0054 | 2.9138 | 2500 | 0.1081 | 0.8503 | 0.8422 | 0.8462 | 0.9834 | | 0.0043 | 3.4965 | 3000 | 0.1210 | 0.8273 | 0.8775 | 0.8516 | 0.9830 | | 0.0049 | 4.0793 | 3500 | 0.1118 | 0.8413 | 0.8590 | 0.8501 | 0.9836 | | 0.0035 | 4.6620 | 4000 | 0.1137 | 0.8465 | 0.8647 | 0.8555 | 0.9837 | | 0.0031 | 5.2448 | 4500 | 0.1150 | 0.8430 | 0.8551 | 0.8490 | 0.9832 | | 0.0027 | 5.8275 | 5000 | 0.1169 | 0.8401 | 0.8590 | 0.8495 | 0.9836 | | 0.0027 | 6.4103 | 5500 | 0.1147 | 0.8517 | 0.8678 | 0.8597 | 0.9847 | | 0.0034 | 6.9930 | 6000 | 0.1163 | 0.8457 | 0.8651 | 0.8553 | 0.9842 | | 0.0024 | 7.5758 | 6500 | 0.1133 | 0.8523 | 0.8652 | 0.8587 | 0.9846 | | 0.0031 | 8.1585 | 7000 | 0.1170 | 0.8399 | 0.8577 | 0.8487 | 0.9834 | | 0.0019 | 8.7413 | 7500 | 0.1243 | 0.8413 | 0.8673 | 0.8541 | 0.9840 | | 0.0019 | 9.3240 | 8000 | 0.1230 | 0.8393 | 0.8726 | 0.8556 | 0.9841 | | 0.002 | 9.9068 | 8500 | 0.1218 | 0.8444 | 0.8549 | 0.8496 | 0.9839 | | 0.002 | 10.4895 | 9000 | 0.1205 | 0.8518 | 0.8651 | 0.8584 | 0.9846 | | 0.0017 | 11.0723 | 9500 | 0.1184 | 0.8643 | 0.8553 | 0.8598 | 0.9846 | | 0.0014 | 11.6550 | 10000 | 0.1316 | 0.8363 | 0.8717 | 0.8536 | 0.9838 | | 0.0016 | 12.2378 | 10500 | 0.1199 | 0.8560 | 0.8660 | 0.8609 | 0.9848 | ### Framework versions - Transformers 4.44.2 - Pytorch 2.1.1+cu121 - Datasets 2.14.5 - Tokenizers 0.19.1
wdfgrfxthb/On-Nutrition-Hunger-or-appetite-By-Barbara-Intermill-fa-updated
wdfgrfxthb
"2024-09-24T23:34:22Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:33:03Z"
--- language: - en --- [![Build Status](http://img2.chinadaily.com.cn/images/202409/24/66f1ff90a3103711c348c73e.jpeg)]() read the full article here : https://datastudio.google.com/embed/s/i2dNZ9_lSGE Source : https://lookerstudio.google.com/embed/s/uuja4QnGi94 Flash News : https://lookerstudio.google.com/embed/s/nCaMUZZ7f7g Biden last Talk : https://datastudio.google.com/embed/s/vleyffYWQoo Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/k2v-FMVI4yw Other Sources : https://datastudio.google.com/embed/s/g92kVt_694o https://datastudio.google.com/embed/s/nNMjpF2Rzmo https://lookerstudio.google.com/embed/s/qc5cupewrq8 https://lookerstudio.google.com/embed/s/txnmomN0EB0 https://datastudio.google.com/embed/s/rILS4PoU9MI https://lookerstudio.google.com/embed/s/nNMjpF2Rzmo https://lookerstudio.google.com/embed/s/uuja4QnGi94 https://datastudio.google.com/embed/s/vagI_SwlB1s https://lookerstudio.google.com/embed/s/pNYOMOr98-g https://lookerstudio.google.com/embed/s/te1NccZJvNE https://datastudio.google.com/embed/s/vwqMZKIT24E https://lookerstudio.google.com/embed/s/oNKiwoGOrOA https://lookerstudio.google.com/embed/s/lrcgoctU72E https://datastudio.google.com/embed/s/uozd1c3kUN0 A scene from the 1978 western comedy "Goin' South" is not easy to forget. At a wild get-together with his old gang, Jack Nicholson's character asks, "Anybody hungry?" A scraggly-looking fellow spits out that he's so hungry, he could "eat a frozen dog." "Well, we'll just go out to the kitchen and see if we got one already froze," Nicholson replies. That would be a fairly accurate description of hunger, say scientists who study such things. Hunger is a physical feeling that drives us to seek food. And it can be uncomfortable if eating is delayed too long. Hunger has varying levels. I might feel more hungry for dinner when my lunch has been smaller than usual. But it's not the same level as being so hungry you can't sleep. Appetite is somewhat different. It is more a desire to eat, sometimes whether we're hungry or not. For example, I might be completely satisfied after eating dinner and still want to eat a bag of popcorn at the movie. That's appetite. Who cares? If we can understand the differences between wanting to eat and needing to eat, perhaps we can make better choices, say experts. Turns out our quest to eat -- as well as our feeling of satisfaction after eating -- is controlled in large part by hormones. Two of the main ones are ghrelin and leptin. When your stomach is needing food, ghrelin sends hunger signals to your brain. Then after you eat and your tummy is satisfied, ghrelin shuts down and lepin tells your brain you are no longer hungry. It's complicated but research has shed some light on how we can help these hormones do their job most efficiently. Want to control your appetite and avoid overeating? Include protein-rich foods with each meal such as eggs, dairy foods, fish, meat, poultry, soy and nuts. Results from several studies published in a 2020 issue of Physiology of Behavior concluded that eating protein "suppresses appetite and decreases ghrelin." When and how much we eat during the day may also influence our appetites, say researchers. A 2022 randomized controlled feeding trial (the best type of study) published in Cell Metabolism looked at how the timing and amount of meals throughout the day affected the hunger and appetite of 30 overweight men and women who wanted to lose weight. Compared to eating fewer calories in the morning and more in the evening, the participants who ate more calories for breakfast and fewer calories in the evening reported significantly less hunger during the day. And tune in to your level of stress, which can cause appetite hormones to go haywire. Rummaging through the cabinet for comfort food at the end of long day might be one signal.....
topooooli/aviny
topooooli
"2024-09-24T23:33:28Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:33:27Z"
Entry not found
xueyj/google-gemma-2b-1727220929
xueyj
"2024-09-24T23:35:44Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:35:29Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
pattara12345/outs
pattara12345
"2024-09-24T23:38:27Z"
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "base_model:biodatlab/whisper-th-medium-combined", "base_model:finetune:biodatlab/whisper-th-medium-combined", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
"2024-09-24T23:35:37Z"
--- library_name: transformers license: apache-2.0 base_model: biodatlab/whisper-th-medium-combined tags: - generated_from_trainer model-index: - name: outs results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # outs This model is a fine-tuned version of [biodatlab/whisper-th-medium-combined](https://huggingface.co/biodatlab/whisper-th-medium-combined) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1967 - Cer: 9.6801 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Cer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.2092 | 1.0 | 537 | 0.1731 | 11.8383 | | 0.1351 | 2.0 | 1074 | 0.2064 | 11.7219 | | 0.056 | 3.0 | 1611 | 0.2102 | 10.7808 | | 0.0429 | 4.0 | 2148 | 0.2010 | 12.1774 | | 0.0313 | 5.0 | 2685 | 0.1967 | 9.6801 | ### Framework versions - Transformers 4.44.2 - Pytorch 2.4.1+cu121 - Datasets 3.0.0 - Tokenizers 0.19.1
FanFierik/SprigPlantar
FanFierik
"2024-09-24T23:35:53Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:35:47Z"
Entry not found
dogssss/Qwen-Qwen1.5-1.8B-1727220950
dogssss
"2024-09-24T23:35:54Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:35:50Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/Calif-Gov-Newsom-Signs-Bill-Banning-Plastic-Bags-At-Grocery-Stores-Into-Law-12-updated
wdfgrfxthb
"2024-09-24T23:35:59Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:35:57Z"
Entry not found
ahmedheakl/asm2asm-deepseek-1.3b-100k-x86-arm-O2
ahmedheakl
"2024-09-25T01:31:48Z"
0
0
null
[ "safetensors", "llama", "region:us" ]
null
"2024-09-24T23:35:58Z"
Entry not found
wdfgrfxthb/In-abortion-fight-Florida-Gov-DeSantis-says-hes-exempt-from-election-interference-law-cd-updated
wdfgrfxthb
"2024-09-24T23:38:15Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:36:54Z"
--- language: - en --- [![Build Status](https://media.zenfs.com/en/the_huffington_post_584/36bc2e7ff9b28f4624244c1627b1147e)]() read the full article here : https://lookerstudio.google.com/embed/s/oNpqPq8RxB0 Source : https://lookerstudio.google.com/embed/s/jCZPTHpffs8 Flash News : https://lookerstudio.google.com/embed/s/nWn2zUVnsTY Biden last Talk : https://datastudio.google.com/embed/s/gyvoIWv8fI0 Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/vL-INTSRUO4 Other Sources : https://datastudio.google.com/embed/s/q5U_-lhOLxU https://lookerstudio.google.com/embed/s/pNYOMOr98-g https://lookerstudio.google.com/embed/s/iwTW5lAWvKM https://lookerstudio.google.com/embed/s/qvAF3DLlkXA https://datastudio.google.com/embed/s/j4FBeV9AGqU https://lookerstudio.google.com/embed/s/kqiIQ1hC76o https://lookerstudio.google.com/embed/s/jLypeiT2Z1M https://lookerstudio.google.com/embed/s/nbbS5frQ_bY https://lookerstudio.google.com/embed/s/uNHf6hABzSw https://datastudio.google.com/embed/s/jsCicuJgMh0 https://lookerstudio.google.com/embed/s/vdG2b9StSd8 https://datastudio.google.com/embed/s/gwNWiyyurDo https://datastudio.google.com/embed/s/q5E-orHtpDc https://lookerstudio.google.com/embed/s/rpw-MogKBY8 Florida will have the chance to vote on whether or not to make abortion legal beyond the six weeks the current law allows. In opposing the state's abortion rights ballot measure, Florida Attorney General Ashley Moody's office is saying she and Gov. Ron DeSantis are exempt from a state law barring them from using their "official authority or influence for the purpose of interfering with an election." That's according to a Monday evening response her office filed to a legal action accusing them of abusing their offices in opposing Amendment 4, which would ensure abortion access in Florida if gets at least 60% of the vote in November. "The executive branch is well within its rights in expressing its concerns about a proposed amendment to the State's governing charter," Moody's legal team wrote. In suing Moody, DeSantis and Jason Weida, secretary of the Agency for Health Care Administration, a Lake Worth attorney had asked the Florida Supreme Court to intervene and stop what he accuses of being illegal government interference. That attorney, Adam Richardson, took issue with how AHCA published a webpage bashing the amendment and then put a television advertisement linking to it. Also mentioned in Richardson's filing is how the Governor's Faith and Community Initiative reached out to religious groups to advertise a call with Moody titled, "Your Legal Rights & Amendment 4's Ramifications." Joining Moody on that call, according to the email in question, was Mat Staver, the founder and chairman of Liberty Counsel, a Christian ministry that has also fought against gay marriage. Richardson said the state's actions "aim to interfere with the people's right to decide whether or not to approve a citizen-initiated proposal to amend their Constitution, free from undue government interference." He cited a law that said state officials can't use their "official authority or influence for the purpose of interfering with an election or a nomination of office or coercing or influencing another person's vote or affecting the result thereof." But Moody's office said there were procedural issues with the lawsuit and that Richardson omitted lines of the law that it maintains exempt Moody, DeSantis and Weida. "That broad exemption for the state's highest-ranking officials accords with the state's 'right to 'speak for itself,'" they wrote citing another case. "The state 'is entitled to say what it wishes,' and to select the views that it wants to express." Moreover, they wrote that the actions weren't election interference but "good government." Richardson "is free to disagree with the content of the webpage, but he has no right to silence (us) from voicing serious concerns about the proposed amendment and the misinformation spread by its proponents," the Monday filing says. Liberty Counsel filed a brief to the court supporting the state earlier Monday, arguing DeSantis and Moody had the First Amendment right to weigh in on the amendment and that the health agency's actions were protected "government speech." "Because the speech at issue is government speech, it is not subject to First Amendment limitations applicable to regulations of private speech," the Liberty Counsel wrote. "The Agency's website and Weida's posts are curated by the government to communicate its views on Amendment 4." Weida had posted on social media advertising the webpage that criticized the amendment. That move was the focus of a different lawsuit that's also still being litigated. Floridians Protecting Freedom, the group behind Florida's abortion rights ballot measure, sued the Agency for Health Care Administration in state court for its webpage and advertisement. There's a virtual hearing for that case scheduled for Wednesday afternoon with Tallahassee-based Circuit Judge Jonathan Sjostrom. "Florida's government has crossed a dangerous line by using public resources to mislead voters and manipulate their choices in the upcoming election," said Michelle Morton, staff attorney for the American Civil Liberties Union of Florida, in a statement announcing that lawsuit. The ACLU of Florida and Southern Legal Counsel are representing Floridians Protecting Freedom in that case. Can't read the document above? Click here.....
xueyj/Qwen-Qwen1.5-1.8B-1727221022
xueyj
"2024-09-24T23:37:16Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:37:02Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/Qwen-Qwen1.5-7B-1727221046
xueyj
"2024-09-24T23:37:30Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:37:26Z"
Entry not found
Krabat/Qwen-Qwen1.5-1.8B-1727221086
Krabat
"2024-09-24T23:38:11Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:38:07Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
SALUTEASD/Qwen-Qwen1.5-1.8B-1727221101
SALUTEASD
"2024-09-24T23:38:35Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
"2024-09-24T23:38:22Z"
--- base_model: Qwen/Qwen1.5-1.8B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
CCLINICON/r2
CCLINICON
"2024-09-24T23:41:33Z"
0
0
null
[ "region:us" ]
null
"2024-09-24T23:39:39Z"
Entry not found
dogssss/Qwen-Qwen1.5-0.5B-1727221216
dogssss
"2024-09-24T23:40:26Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
"2024-09-24T23:40:17Z"
--- base_model: Qwen/Qwen1.5-0.5B library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
wdfgrfxthb/In-abortion-fight-Florida-Gov-DeSantis-says-hes-exempt-from-election-interference-law-g4-updated
wdfgrfxthb
"2024-09-24T23:41:38Z"
0
0
null
[ "en", "region:us" ]
null
"2024-09-24T23:40:21Z"
--- language: - en --- [![Build Status](https://cdn.abcotvs.com/dip/images/15345869_092324-wpvi-kings-highway-improvement-trish-430pm-CC-vid.jpg?w=1600)]() read the full article here : https://lookerstudio.google.com/embed/s/vdG2b9StSd8 Source : https://lookerstudio.google.com/embed/s/gdzhtJq0sIA Flash News : https://datastudio.google.com/embed/s/r1r2qy_Z0Pk Biden last Talk : https://datastudio.google.com/embed/s/s72el1tNXnY Russian Ukrain Breaking News : https://lookerstudio.google.com/embed/s/kaltQLA_KSs Other Sources : https://lookerstudio.google.com/embed/s/vvJd5WugEWw https://datastudio.google.com/embed/s/pUgJav5zIgI https://lookerstudio.google.com/embed/s/g92kVt_694o https://datastudio.google.com/embed/s/mukAJsNg1y4 https://lookerstudio.google.com/embed/s/oESW8NiQ5DU https://datastudio.google.com/embed/s/sgN1oYeql4A https://lookerstudio.google.com/embed/s/i0uzWrWG4Vg https://lookerstudio.google.com/embed/s/uiXSFOYm1UU https://lookerstudio.google.com/embed/s/lrcgoctU72E https://datastudio.google.com/embed/s/nCaMUZZ7f7g https://datastudio.google.com/embed/s/iwTW5lAWvKM https://datastudio.google.com/embed/s/rs0m_UgZb3k https://datastudio.google.com/embed/s/hy8cvKi8tDE https://datastudio.google.com/embed/s/jlKGUqDLiZc Florida will have the chance to vote on whether or not to make abortion legal beyond the six weeks the current law allows. In opposing the state's abortion rights ballot measure, Florida Attorney General Ashley Moody's office is saying she and Gov. Ron DeSantis are exempt from a state law barring them from using their "official authority or influence for the purpose of interfering with an election." That's according to a Monday evening response her office filed to a legal action accusing them of abusing their offices in opposing Amendment 4, which would ensure abortion access in Florida if gets at least 60% of the vote in November. "The executive branch is well within its rights in expressing its concerns about a proposed amendment to the State's governing charter," Moody's legal team wrote. In suing Moody, DeSantis and Jason Weida, secretary of the Agency for Health Care Administration, a Lake Worth attorney had asked the Florida Supreme Court to intervene and stop what he accuses of being illegal government interference. That attorney, Adam Richardson, took issue with how AHCA published a webpage bashing the amendment and then put a television advertisement linking to it. Also mentioned in Richardson's filing is how the Governor's Faith and Community Initiative reached out to religious groups to advertise a call with Moody titled, "Your Legal Rights & Amendment 4's Ramifications." Joining Moody on that call, according to the email in question, was Mat Staver, the founder and chairman of Liberty Counsel, a Christian ministry that has also fought against gay marriage. Richardson said the state's actions "aim to interfere with the people's right to decide whether or not to approve a citizen-initiated proposal to amend their Constitution, free from undue government interference." He cited a law that said state officials can't use their "official authority or influence for the purpose of interfering with an election or a nomination of office or coercing or influencing another person's vote or affecting the result thereof." But Moody's office said there were procedural issues with the lawsuit and that Richardson omitted lines of the law that it maintains exempt Moody, DeSantis and Weida. "That broad exemption for the state's highest-ranking officials accords with the state's 'right to 'speak for itself,'" they wrote citing another case. "The state 'is entitled to say what it wishes,' and to select the views that it wants to express." Moreover, they wrote that the actions weren't election interference but "good government." Richardson "is free to disagree with the content of the webpage, but he has no right to silence (us) from voicing serious concerns about the proposed amendment and the misinformation spread by its proponents," the Monday filing says. Liberty Counsel filed a brief to the court supporting the state earlier Monday, arguing DeSantis and Moody had the First Amendment right to weigh in on the amendment and that the health agency's actions were protected "government speech." "Because the speech at issue is government speech, it is not subject to First Amendment limitations applicable to regulations of private speech," the Liberty Counsel wrote. "The Agency's website and Weida's posts are curated by the government to communicate its views on Amendment 4." Weida had posted on social media advertising the webpage that criticized the amendment. That move was the focus of a different lawsuit that's also still being litigated. Floridians Protecting Freedom, the group behind Florida's abortion rights ballot measure, sued the Agency for Health Care Administration in state court for its webpage and advertisement. There's a virtual hearing for that case scheduled for Wednesday afternoon with Tallahassee-based Circuit Judge Jonathan Sjostrom. "Florida's government has crossed a dangerous line by using public resources to mislead voters and manipulate their choices in the upcoming election," said Michelle Morton, staff attorney for the American Civil Liberties Union of Florida, in a statement announcing that lawsuit. The ACLU of Florida and Southern Legal Counsel are representing Floridians Protecting Freedom in that case. Can't read the document above? Click here.....
Mottzerella/Llama-3-8B-Instruct-Finance-RAG
Mottzerella
"2024-09-24T23:42:30Z"
0
0
null
[ "base_model:meta-llama/Meta-Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3.1-8B-Instruct", "region:us" ]
null
"2024-09-24T23:40:37Z"
--- base_model: - meta-llama/Meta-Llama-3.1-8B-Instruct ---
Krabat/google-gemma-2b-1727221251
Krabat
"2024-09-24T23:40:56Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:40:51Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
xueyj/google-gemma-2b-1727221256
xueyj
"2024-09-24T23:41:21Z"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-2b", "base_model:adapter:google/gemma-2b", "region:us" ]
null
"2024-09-24T23:40:56Z"
--- base_model: google/gemma-2b library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.0
Grayx/fiufiu_12
Grayx
"2024-09-24T23:43:53Z"
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-09-24T23:41:13Z"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]