Dataset Viewer
Auto-converted to Parquet Duplicate
modelId
stringlengths
4
81
tags
sequence
pipeline_tag
stringclasses
17 values
config
dict
downloads
int64
0
59.7M
first_commit
timestamp[ns, tz=UTC]
card
stringlengths
51
438k
embedding
sequence
albert-base-v1
[ "pytorch", "tf", "safetensors", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "exbert", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
38,156
2019-12-20T12:28:51
--- tags: - exbert language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT Base v1 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github...
[ -0.01628965139389038, 0.004767254460602999, -0.02520707994699478, 0.06764831393957138, 0.03774816170334816, 0.02252449281513691, -0.016135185956954956, -0.03889353945851326, -0.03518366441130638, 0.058156222105026245, 0.02774147316813469, -0.0001626275625312701, 0.00031815734109841287, 0.0...
albert-base-v2
[ "pytorch", "tf", "jax", "rust", "safetensors", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
4,785,283
2019-11-04T16:00:52
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT Base v2 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-rese...
[ -0.017371734604239464, 0.0015233331359922886, -0.02679620310664177, 0.06894268840551376, 0.035669613629579544, 0.021609894931316376, -0.017564035952091217, -0.037039611488580704, -0.03948637843132019, 0.05711784213781357, 0.0304550938308239, 0.0007140071247704327, 0.004628319758921862, 0.0...
albert-large-v1
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
687
2019-12-20T12:28:51
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT Large v1 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-res...
[ -0.018121063709259033, 0.002527142409235239, -0.026467183604836464, 0.06899815797805786, 0.035836223512887955, 0.019501175731420517, -0.017295489087700844, -0.03910917416214943, -0.03756653517484665, 0.05714162811636925, 0.030745433643460274, 0.0013536266051232815, 0.004762119147926569, 0....
albert-large-v2
[ "pytorch", "tf", "safetensors", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
26,792
2019-11-04T16:00:53
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT Large v2 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-res...
[ -0.017891714349389076, 0.002103835577145219, -0.025990311056375504, 0.0689493864774704, 0.036186132580041885, 0.019427411258220673, -0.017666058614850044, -0.03851858526468277, -0.03691915050148964, 0.0569012388586998, 0.03065728396177292, 0.002100828569382429, 0.00448606489226222, 0.03260...
albert-xlarge-v1
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
341
2019-12-20T12:28:51
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT XLarge v1 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-re...
[ -0.017884569242596626, 0.006399250589311123, -0.020775070413947105, 0.06745092570781708, 0.03645424172282219, 0.019362691789865494, -0.020082443952560425, -0.04499386250972748, -0.031038735061883926, 0.05520254373550415, 0.03186742216348648, -0.0028629458975046873, 0.0012352790217846632, 0...
albert-xlarge-v2
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
2,973
2019-11-04T16:00:53
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT XLarge v2 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-re...
[ -0.017797056585550308, 0.005246325396001339, -0.020524784922599792, 0.06777287274599075, 0.03693637624382973, 0.019691655412316322, -0.020058730617165565, -0.043487515300512314, -0.03131895139813423, 0.055198222398757935, 0.031354181468486786, -0.0017217992572113872, 0.0013881685445085168, ...
albert-xxlarge-v1
[ "pytorch", "tf", "albert", "fill-mask", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1909.11942", "transformers", "license:apache-2.0", "autotrain_compatible", "has_space" ]
fill-mask
{ "architectures": [ "AlbertForMaskedLM" ], "model_type": "albert", "task_specific_params": { "conversational": { "max_length": null }, "summarization": { "early_stopping": null, "length_penalty": null, "max_length": null, "min_length": null, "no_repeat_ngram_...
7,091
2019-12-20T12:28:51
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia --- # ALBERT XXLarge v1 Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1909.11942) and first released in [this repository](https://github.com/google-r...
[ -0.02078196220099926, 0.006141583435237408, -0.0204166267067194, 0.0693066269159317, 0.036629922688007355, 0.01839315891265869, -0.018037041649222374, -0.04531469568610191, -0.03084694966673851, 0.05282839015126228, 0.03224258869886398, -0.0016929764533415437, 0.0014640215085819364, 0.0344...
albert-xxlarge-v2
["pytorch","tf","safetensors","albert","fill-mask","en","dataset:bookcorpus","dataset:wikipedia","ar(...TRUNCATED)
fill-mask
{"architectures":["AlbertForMaskedLM"],"model_type":"albert","task_specific_params":{"conversational(...TRUNCATED)
42,640
2019-11-04T16:00:52
"---\ntags:\n- exbert\nlanguage: en\nlicense: apache-2.0\ndatasets:\n- bookcorpus\n- wikipedia\n---\(...TRUNCATED)
[-0.019903937354683876,0.008878600783646107,-0.02040528692305088,0.06702258437871933,0.0376590751111(...TRUNCATED)
bert-base-cased
["pytorch","tf","jax","safetensors","bert","fill-mask","en","dataset:bookcorpus","dataset:wikipedia"(...TRUNCATED)
fill-mask
{"architectures":["BertForMaskedLM"],"model_type":"bert","task_specific_params":{"conversational":{"(...TRUNCATED)
8,621,271
2018-11-14T23:35:08
"---\nlanguage: en\ntags:\n- exbert\nlicense: apache-2.0\ndatasets:\n- bookcorpus\n- wikipedia\n---\(...TRUNCATED)
[-0.005537884775549173,0.0068740639835596085,-0.01787860319018364,0.06503400951623917,0.026847530156(...TRUNCATED)
bert-base-chinese
["pytorch","tf","jax","safetensors","bert","fill-mask","zh","arxiv:1810.04805","transformers","autot(...TRUNCATED)
fill-mask
{"architectures":["BertForMaskedLM"],"model_type":"bert","task_specific_params":{"conversational":{"(...TRUNCATED)
3,377,486
2018-11-14T23:35:08
"---\nlanguage: zh\n---\n\n# Bert-base-chinese\n\n## Table of Contents\n- [Model Details](#model-det(...TRUNCATED)
[-0.027363037690520287,-0.012506258673965931,-0.0041877892799675465,0.06901980191469193,0.0159092675(...TRUNCATED)
End of preview. Expand in Data Studio

Dataset Card for "card_with_first_commit_embedded"

More Information needed

Downloads last month
2