Paper page TinyStories: How Small Can Language Models Be and Still Speak Coherent English?

LLMs vs SLMs: The Differences in Large & Small Language Models

small language models

As toxicity is culturally sensitive, attempting to find equivalents in a largely multilingual setting constitutes a challenge when starting from one source language. To address this issue, translators were allowed to forgo translating some of the source items and add more culturally relevant items. However, as we increase the model capacity and the computational cost per update, the propensity for low or very low-resource languages to overfit increases, thus causing performance to deteriorate.

Since the community creates these characters, false results, called hallucinations, are frequently generated. When you begin chatting with the various characters, it’s important to consider where they originate from and expect that most, if not all, of what they say is made up. While you can enable your characters to generate images, they do not belong to the same class as other AI art generators, primarily because it was created mainly as a text generator.

For other datasets, while there might be visual differences in performance with and without instruction fine-tuning, these differences aren’t statistically significant based on the p-values. For many datasets, instruction fine-tuning improves performances when compared to not fine-tuning (e.g., agnews, ethos, imdb, trec, yelp, and youtube). This is evident from the graphical representation and the significant p-values from the ANCOVA. Datasets like bbcnews, youtube, and sms show a decrease in performance when instruction fine-tuning is applied, but ANCOVA tells us that it is not significant. Figure 3 visually compares the impact of instruction-tuning and performance metrics (Acc/F1) across various datasets. From our analysis, 10 of 15 datasets show p-values exceeding 0.05, suggesting no significant link between Acc/F1 scores and model size.

By adhering to these principles, you can navigate challenges effectively and achieve optimal project results. What are the typical hardware requirements for deploying and running Small Language Models? One of the key benefits of Small Language Models is their reduced hardware requirements compared to Large Language Models. Typically, SLMs can be run on standard laptop or desktop computers, often requiring only a few gigabytes of RAM and basic GPU acceleration. This makes them much more accessible for deployment in resource-constrained environments, edge devices, or personal computing setups, where the computational and memory demands of large models would be prohibitive.

small language models

You can foun additiona information about ai customer service and artificial intelligence and NLP. The entertainment industry is undergoing a transformative shift, with SLMs playing a central role in reshaping creative processes and enhancing user engagement. But, if you all can pull it off, good luck with the idea as the concept of the shared resource isn’t bad, so ill shut up now. Another use case might be data parsing/annotating, where you can prompt an SLM to read from files/spreadsheets. It can then (a) rewrite the information in your data in the format of your choice, and (b) add annotations and infer metadata attributes for your data. To sum it up, no matter which model architecture we look at, the choice of scoring function doesn’t seem to affect more than another.

From LLaMA to Claude 3 to Command-R and more, companies have been releasing their own rivals to GPT-4, OpenAI’s latest large multimodal model. As we noticed difference in classification performances under different scoring functions but none could lead to a clear winner, could’nt juge really how well models performed. So we decided to take the mean of these scores to have a more robust evaluation of the model’s performance. For both encoder-decoder and decoder-only models, values are above the standard 0.05 by a large margin. In prompt-based classification, using a verbalizer mapping tokens to class labels is crucial for accurate classification.

Moreover, we observe that languages within the same family are highly similar in their choice of experts (that is, the late decoder MoE layers are language-specific). This is particularly the case for the Arabic dialects (the six rows and columns in the top-left corner), languages in the Benue–Congo subgrouping, as well as languages in the Devanagari script. By contrast, the early decoder MoE layers (Fig. 1c) seem to be less language-specific. The late encoder MoE layers are particularly language-agnostic in how they route tokens as can be attested by the uniform heat map in Fig.

However, building and implementing an effective SLM requires expertise, resources, and a strategic approach. After successfully downloading the pre-trained model, you will need to load it into your Python environment. Pay close attention to detail during the loading process to avoid common pitfalls. Depending on the library and framework you’re using, specific functions or classes are available for loading models. For instance, TensorFlow provides the tf.saved_model.load() function for this purpose.

These frameworks epitomize the evolving landscape of AI customization, where developers are empowered to create SLMs tailored to specific needs and datasets. With these tools at their disposal, organizations across industries can harness the transformative potential of bespoke language models, driving innovation and unlocking new opportunities in the realm of AI-driven solutions. Small language models are essentially more streamlined versions of LLMs, in regards to the size of their neural networks, and simpler architectures. Compared to LLMs, SLMs have fewer parameters and don’t need as much data and time to be trained — think minutes or a few hours of training time, versus many hours to even days to train a LLM.

That would theoretically not only save money in the long run but also require far less energy in aggregate, dramatically decreasing AI’s environmental footprint. AI models like Phi-3 may be a step toward that future if the benchmark results hold up to scrutiny. When trained on cleaner and less noisy data, smaller models can potentially encapsulate comparable intelligence in significantly fewer parameters. While large language models certainly hold a place in the AI landscape, the momentum appears to be favoring compact, specialized models. Hugging Face stands at the forefront of democratizing AI with its comprehensive Hub.

Tabnine offers three plans, including the Starter plan, which is completely free. For more features, the Pro plan offers collaboration for up to 100 users, whole line and full-function code completions, and natural language to code completions for a very affordable $12 per month per user. An excellent feature of Tabnine is its ability to adapt to the individual user’s coding style. Because of this, it can predict and suggest lines of code based on context, allowing users to streamline repetitive tasks to produce high-quality code. Tabnine’s deep learning algorithms also enable it to offer high-quality suggestions for multiple coding languages, so no matter what type of project you’re working on, Tabnine has a solution. Tabnine is an AI-driven coding assistant that boosts productivity by enabling developers to write code quickly and effectively.

On the flip side, the increased efficiency and agility of SLMs may translate to slightly reduced language processing abilities, depending on the benchmarks the model is being measured against. I wonder if we could shard the dataset somehow and have the community do federated data cleansing? The choice and assumptions of the statistical tools could influence the results.

Developers looking to improve their code quality and security through automated code reviews and static code analysis will love Codiga. It supports multiple programming languages, offers custom rule sets, and integrates with all major IDEs, so it’s a great tool for fixing code errors and identifying security vulnerabilities. That, on top of code snippet sharing and management features, makes Codiga an excellent choice.

Collections including this paper

We called small language models, models within the size range 77M to 3B parameters. These models are comparatively smaller, ranging from 13 to 156 times less in parameter count than our largest model, Falcon 40B111We do not test Falcon 180B, as it was not released during our experiments. Moreover, at the time our study was conducted, TinyStories (Eldan and Li, 2023) models, which are on an even smaller scale, starting at 1M parameters. With the free plan, new users or casual coders can have 500 monthly autocompletions, 20 messages or commands, personalization for small codebases, and large language model (LLM) support. It has unlimited autocompletions, messages, commands, and personalizations for any codebase size and multiple LLM choices.

  • To ensure low-resource languages are well-represented in the vocabulary, we downsampled high-resource and upsampled low-resource languages with a sampling temperature of five (ref. 10).
  • An ANCOVA is made to quantify the impact of instruction-tuning on each architecture (encoder-decoder/decoder-only) while statistically controlling for the effect of the model size feature.
  • It can automatically grab the proper selectors of your module and apply the exact CSS of your request to them.
  • Some people found the earlier Llama 2 model — released less than a year ago — to be “a little stiff and sanctimonious sometimes in not responding to what were often perfectly innocuous or innocent prompts and questions,” he said.

WordPress devs might be interested in our new feature for our Divi called Divi Snippets. It allows developers to save and manage their most used code snippets, including HTML, Javascript, CSS, and collections of CSS parameters and rules. This is a perfect companion tool for WordPress developers using some https://chat.openai.com/ of the best AI coding assistants to improve the quality of their work. SinCode offers a free plan with limited access to basic features, such as Marve (GPT 3.5) and limited image generation. Word credits can be purchased for $4.50 per 3,000 words, including 10 images, GPT-4, GPT 3.5 Turbo, and Marve Chat.

Using them creates efficiencies at every stage of development, no matter what type of project you are working on. Many of the best development teams have already switched to many of the solutions below. Two popular platforms, Shopify and Etsy, have the potential to turn those dreams into reality. Buckle up because we’re diving into Shopify vs. Etsy to see which fits your unique business goals! As previously mentioned, most of the output is likely false, so checking what it gives you is important. After playing with the Translator bot, we can say that it is mostly accurate and had no trouble translating a simple sentence into Urdu, the primary language spoken in Pakistan.

Example Applications Where Small Language Models Shine

This intentional design choice enhances computational efficiency and task-specific effectiveness without sacrificing linguistic comprehension and generation capabilities. Generally, researchers agree that language models with fewer than 100 million parameters fall under the “small” category, although this classification can differ. Some specialists consider models with parameter counts ranging from one million to 10 million as small, especially when compared to contemporary large models, which may have hundreds of billions of parameters.

small language models

In addition, they want to probe the ability of large language models to exhibit spatial awareness and see how this could aid language-based navigation. Current approaches often utilize multiple hand-crafted machine-learning models to tackle different parts of the task, which require a great deal of human effort and expertise to build. These methods, which use visual representations to directly make navigation decisions, demand massive amounts of visual data for training, which are often hard to come by. Language identification is a challenging task in which numerous failure modes exist, often exacerbated by the gaps between the clean data on which LID models are trained and noisy data on which LID models are applied. In other words, LID models trained in a supervised manner on fluently written sentences may have difficulty identifying grammatically incorrect and incomplete strings extracted from the web. Furthermore, models can easily learn spurious correlations that are not meaningful for the task itself.

Cohere’s developer-friendly platform enables users to construct SLMs remarkably easily, drawing from either their proprietary training data or imported custom datasets. Offering options with as few as 1 million parameters, Cohere ensures flexibility without compromising on end-to-end privacy compliance. With Cohere, developers can seamlessly navigate the complexities of SLM construction while prioritizing data privacy. Follow these simple steps to unlock the versatile and efficient capabilities of small language models, rendering them invaluable for a wide range of language processing tasks. With the correct setup and optimization, you’ll be empowered to tackle NLP challenges effectively and achieve your desired outcomes. To start the process of running a language model on your local CPU, it’s essential to establish the right environment.

But it’s often hard to predict which characteristics of small models will also appear in larger ones. Eldan now had a procedure for churning out training data on demand, but he had no idea how many stories he’d need to train a functional model, or how big that model would need to be. That’s when he teamed up with Yuanzhi Li, a machine learning researcher at Microsoft and Carnegie Mellon University, to try different possibilities, taking advantage of the fact that small models could be trained very quickly.

Lamda used a decoder-only transformer language model and was pre-trained on a large corpus of text. In 2022, LaMDA gained widespread attention when then-Google engineer Blake Lemoine went public with claims that the program was sentient. The bot was released in August 2023 and has garnered more than 45 million users. Lists are based on professional translations from English, which were then heuristically adapted by linguists to better serve the target language.

This targeted training allows them to achieve high accuracy on relevant tasks while remaining computationally frugal. With significantly fewer parameters (ranging from millions to a few billion), they require less computational power, making them ideal for deployment on mobile devices and resource-constrained environments. In the context of artificial intelligence and natural language processing, SLM can stand for ‘Small Language Model’. The label “small” in this context refers to a) the size of the model’s neural network, b) the number of parameters and c) the volume of data the model is trained on. There are several implementations that can run on a single GPU, and over 5 billion parameters, including Google Gemini Nano, Microsoft’s Orca-2–7b, and Orca-2–13b, Meta’s Llama-2–13b and others. Android Studio Bot is the best AI coding assistant for those creating Android apps and wanting to boost their productivity.

small language models

We want to see how small models perform in this zero-shot text classification and determine what makes them do well with specific data. We are comparing how small and big models work with zero-shot prompting on various data sets to understand if we can get good results with less resources. Large Language Models (LLMs) have been massively favored over smaller models to solve tasks through prompting (Brown et al., 2020; Hoffmann et al., 2022; OpenAI, 2023; Chowdhery et al., 2022) in a zero-shot setting. However, while their utility is extensive, they come with challenges – they are resource-intensive, costly to employ, and their performances are not always warranted for every task (Nityasya et al., 2021). Bigger models (Kaplan et al., 2020; Hoffmann et al., 2022) were built, always sophisticated datasets were necessary (Zhang et al., 2023) to achieve claimed performances.

Before feeding your data into the language model, it’s imperative to preprocess it effectively. This may involve tokenization, stop word removal, or other data cleaning techniques. Since each language model may have specific requirements for input data formatting, consulting the documentation for your chosen model is essential to ensure compatibility. According to Microsoft, the efficiency of the transformer-based Phi-2 makes it an ideal choice for researchers who want to improve safety, interpretability and ethical development of AI models.

The goal is to use the learned probability distribution of natural language for generating a sequence of phrases that are most likely to occur based on the available contextual knowledge, which includes user prompt queries. In comparison, the largest model yet released in Meta’s Llama 3 family includes 70 billion parameters (with a 400 billion version on the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused on making smaller AI language models as capable as larger ones were a few years ago. In the world of AI, what might be called “small language models” have been growing in popularity recently because they can be run on a local device instead of requiring data center-grade computers in the cloud. On Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly on a smartphone.

In this section, we examine how we can use Sparsely Gated Mixture of Experts models2,3,4,5,6,7 to achieve a more optimal trade-off between cross-lingual transfer and interference and improve performance for low-resource languages. We hypothesize that added toxicity may be because of the presence of toxicity in the training data and used our detectors to estimate, more specifically, unbalanced toxicity in the bitext data. We find that estimated levels of unbalanced toxicity vary from one corpus of bitext to the next and that unbalanced toxicity can be greatly attributed to misaligned bitext. In other words, training with this misaligned bitext could encourage mistranslations with added toxicity.

Depending on the number of concurrent users accessing an LLM, the model inference tends to slow down. They may lack holistic contextual information from all multiple knowledge domains but are likely to excel in their chosen domain. Language models are AI computational models that can generate natural human language.

There are several fine-tuned versions of Palm, including Med-Palm 2 for life sciences and medical information as well as Sec-Palm for cybersecurity deployments to speed up threat analysis. But one disadvantage is that their method naturally loses some information that would be captured by vision-based models, such as depth information. To streamline the process, the researchers designed templates so observation information is presented to the model in a standard form — as a series of choices the robot can make based on its surroundings. The model repeats these processes to generate a trajectory that guides the robot to its goal, one step at a time. The large language model outputs a caption of the scene the robot should see after completing that step. This is used to update the trajectory history so the robot can keep track of where it has been.

These architectures are designed to balance performance, efficiency, and accessibility. In the same way as architecture, we quantified the impact of instruction-tuning on performances while controlling the number of parameters. Figure 1 presents the relationship between the number of parameters and the performance in terms of Acc/F1 scores across various datasets. We follow Brown et al. (2020) to craft simple prompts while ensuring domain relevance.

The platform generates code, finds relevant resources, teaches best practices, and saves time. Although the bot is still in the developmental stage, it’s already proven an excellent tool for developers of all skill levels. CodeWP is an AI-powered, cloud-based WordPress code generator designed to simplify the coding process for WordPress developers across all skill levels. This platform can rapidly generate valid code for tasks such as creating custom post types, developing plugins, and extending the core function of your favorite WordPress products.

DSM languages tend to support higher-level abstractions than General-purpose modeling languages, so they require less effort and fewer low-level details to specify a given system. Algebraic Modeling Languages (AML) are high-level programming languages for describing and solving high complexity problems for large scale mathematical Chat GPT computation (i.e. large scale optimization type problems). One particular advantage of AMLs like AIMMS, AMPL, GAMS, Gekko, Mosel, OPL and OptimJ is the similarity of its syntax to the mathematical notation of optimization problems. The algebraic formulation of a model does not contain any hints how to process it.

We applied threshold optimization so that when the confidence of a classifier is low, the corresponding language is not considered for the final decision. A sentence was filtered out if none of the classifiers surpassed its threshold. Second, we built a multiclass classifier using softmax over all possible languages.

  • Included in it are models that paved the way for today’s leaders as well as those that could have a significant effect in the future.
  • LLMs are black box AI systems that use deep learning on extremely large datasets to understand and generate new text.
  • The Claude LLM focuses on constitutional AI, which shapes AI outputs guided by a set of principles that help the AI assistant it powers helpful, harmless and accurate.
  • Building an AI-powered dynamic pricing system involves a systematic approach that integrates advanced technologies to optimize pricing strategies and enhance competitiveness.
  • We did not attempt to optimize the architecture and parameters of the bilingual NMT systems to the characteristics of each language pair but used the same architecture for all.

Someday, you may want your home robot to carry a load of dirty clothes downstairs and deposit them in the washing machine in the far-left corner of the basement. The robot will need to combine your instructions with its visual observations to determine the steps it should take to complete this task. In the initial release of the Toxicity-200 lists, the average number of items in a toxicity detection list was 271 entries, whereas the median number of entries was 143. First, we used a combination of multiple binary classifiers in which the final decision was obtained by selecting the language with the highest score after applying a threshold.

Some of the largest language models today, like Google’s PaLM 2, have hundreds of billions of parameters. OpenAI’s GPT-4 is rumored to have over a trillion parameters but spread over eight 220-billion parameter models in a mixture-of-experts configuration. Both models require heavy-duty data center GPUs (and supporting systems) to run properly.

For example, Efficient Transformers have become a popular small language model architecture employing various techniques like knowledge distillation during training to improve efficiency. Relative to baseline Transformer models, Efficient Transformers achieve similar language task performance with over 80% fewer parameters. Effective architecture decisions amplify the ability companies can extract from small language models of limited scale. Small language models can capture much of this broad competency during pretraining despite having limited parameter budgets. Specialization phases then afford refinement towards specific applications without needing to expand model scale. Overall, transfer learning greatly improves data efficiency in training small language models.

Embedding were created for the answers generated by the SLM and GPT-3.5 and the cosine distance was used to determine the similarity of the answers from the two models. Meta even considered acquiring the publisher Simon & Schuster in a bid to get more data to train its models, The New York Times reported last month. Page Builders gained prominence at a time when designing a website with WordPress entailed knowing HTML, CSS, and some PHP. If you’d allow us to say it, page builders like Divi were a bit of a reassurance for WordPress users…. The best AI coding assistants are, hands down, Github Copilot, Divi AI, and Tabnine.

Relation classification tasks are also included using datasets like semeval (Hendrickx et al., 2010). While previous work focused on new methods to make language models better zero-shot learners, we want insight into model features and how well they perform. Fei et al. (2022) enhances zero-shot classification by segmenting input texts and leveraging class-specific prompts. While Meng et al. (2020) proposed a strategy that employs label names combined with self-training tailored for zero-shot classification. Many methods necessitate an unlabeled dataset or a knowledge base to extract pertinent topic words and facilitate self-training.

It includes algorithms that inform AI models when there’s an error so it can learn from it. The best AI coding assistants have a few things in common, including the ability to generate code, spot within small language models code, complete snippets automatically, and support most major IDEs. The Free plan comes with 100 free actions per month, 1 project, some chat and generation functionality, and community support.

The Pro plan adds 10,000 actions, 4 projects, and 28+ plugin-specific AI models for $28 monthly. Finally, the Agency plan is the most robust, with unlimited actions, 3 team members, unlimited projects, and custom AI models for an affordable $68 monthly. Replit is a powerful tool that allows you to speed up the coding process through artificial intelligence. Those who are learning how to code or want to work in a collaborative environment from anywhere will find Replit a worthy companion.

Although the intent of this declaration was to limit censorship and allow for information and ideas to flow without interference, much of the internet today remains inaccessible to many due to language barriers. Our effort was designed to contribute one solution to help alter this status quo. The framework states the ability to represent the domain as domain appropriateness. The statement appropriateness can be a bit vague, but in this particular context it means able to express.

Overall, a sample of 55 language directions were evaluated, including 8 into English, 27 out of English, and 20 other direct language directions. The overall mean of calibrated XSTS scores was 4.26, with 38/55 directions scoring over 4.0 (that is, high quality) and 52/56 directions scoring over 3.0. The language used is appropriate for the organizational context, e.g. that the language is standardized within the organization, or that it is supported by tools that are chosen as standard in the organization. To ensure that the domain actually modelled is usable for analyzing and further processing, the language has to ensure that it is possible to reason in an automatic way. Another advantage by formalizing is the ability to discover errors in an early stage.

These captions are combined with language-based instructions and fed into a large language model, which decides what navigation step the robot should take next. But such models take text-based inputs and can’t process visual data from a robot’s camera. (A previous detector quality analysis showed that a higher precision was reached in this situation). We added this toxicity filtering procedure as an option to the filtering process and experimented with or without it for comparison.

How can LeewayHertz help you build powerful small language models?

The Pro plan increases the GTP-3.5 generations to 1,000,000 and adds 100,000 GTP-4 tokens for $9 monthly. Finally, the Advanced plan provides a whopping 300,000 GPT-4 tokens, 2 million 3.5 tokens, customizable data dashboards, and connections to outside data sources for $19 monthly. SQLAI.ai is best suited for many users, including beginners, experienced web developers, and data analysts. It is designed to boost SQL productivity and proficiency, offering AI-powered query generation, explanation, and optimization features.

Performance configuration was also enabled for efficient adaptation of pre-trained models. Finally, training arguments were used for defining particulars of the training process and the trainer was passed parameters, data, and constraints. BERT is a transformer-based model that can convert sequences of data to other sequences of data. BERT’s architecture is a stack of transformer encoders and features 342 million parameters. BERT was pre-trained on a large corpus of data then fine-tuned to perform specific tasks along with natural language inference and sentence text similarity. It was used to improve query understanding in the 2019 iteration of Google search.

They are gaining popularity and relevance in various applications especially with regards to sustainability and amount of data needed for training. From the hardware point of view, it is cheaper to run i.e., SLMs require less computational power and memory and it is suitable for on-premises and on-device deployments making it more secure. We then compared NLLB-200 with a few other state-of-the-art models, such as Deepnet42 and M2M-100 (ref. 1), to report scores for 87 languages against FLORES-101. Overall, the results show that NLLB-200 improves on state-of-the-art systems by a notable margin despite supporting 200 languages, or twice as many languages (and more than 30,000 additional directions) compared with any previous work.

Figure 4 visually compares the impact of instruction-tuning and performance metrics (Acc/F1) for the two architectures. We use ANCOVA to test whether the means of our ACC/F1 scores are equal across modalities of instruction tuning while statistically controlling the effect of the number of parameters. On one hand, 7 out of 15 datasets, namely agnews, bbcnews, chemprot, semeval, sms, spouse, and youtube, show p-values bellow 0.05, suggesting there the architecture has a significant impact. Using ANCOVA, we measure the impact of the architecture choice on Acc/F1 scores, while controlling the effect of the model size variable.

Microsoft unveils a new small language model – MarTech

Microsoft unveils a new small language model.

Posted: Fri, 26 Apr 2024 07:00:00 GMT [source]

In conclusion, while many datasets do not show a direct relationship between larger model sizes and improved performance, datasets like cdr, ethos, and imdb do. Overall, the variance in the correlation coefficient across datasets suggests that model size isn’t the sole determinant of performance. Instruction-tuning refers to the strategy for fine-tuning a language model on instruction datasets (Longpre et al., 2023). Please note that we used GPT-3.5 to generate questions and answers from the training data. The model that we fine-tuned is Llama-2–13b-chat-hf has only 13 billion parameters while GPT-3.5 has 175 billion.

Building an enterprise AI solution in logistics involves leveraging advanced technologies to automate processes, gain insights, and make data-driven decisions within logistics operations. Our comprehensive support and maintenance services are designed to uphold the peak performance of your SLM. This includes ongoing monitoring, adaptation to evolving data and use cases, prompt bug fixes, and regular software updates. The broad spectrum of applications highlights the adaptability and immense potential of Small Language Models, enabling businesses to harness their capabilities across industries and diverse use cases. ArXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Additionally, SLMs can be customized to meet an organization’s specific requirements for security and privacy.

We prompt various language models using 4 different scoring functions (see Section 3.4.2) to classify sentences and report accuracy and F1 scores for each triple model-datasets-scoring function. In the dynamic landscape of NLP, small language models serve as catalysts for innovation, democratizing access to advanced language processing tools and fostering inclusivity within the field. Their potential to empower diverse communities and streamline development processes holds promise for driving impactful advancements across numerous sectors, from education to healthcare and beyond.

If you need some assistance, check out the character book, which gives you a wealth of information to help you create your AI characters. One of the best features of Character AI is the ability to create your own chatbot to interact with. The first step is clicking the create button located in the navigation bar on the left-hand side of the interface. First and foremost, it’s a great way to dialogue with different characters, giving you different perspectives. You can chat with Elon Musk, Edward Cullen from the popular Twilight books, or even Taylor Swift.