7.2 C
Canberra
Thursday, October 23, 2025

What We Realized from a 12 months of Constructing with LLMs (Half III): Technique – O’Reilly


We beforehand shared our insights on the techniques we’ve honed whereas working LLM purposes. Ways are granular: they’re the precise actions employed to attain particular targets. We additionally shared our perspective on operations: the higher-level processes in place to assist tactical work to attain targets.


Be taught sooner. Dig deeper. See farther.

However the place do these targets come from? That’s the area of technique. Technique solutions the “what” and “why” questions behind the “how” of techniques and operations.

We offer our opinionated takes, similar to “no GPUs earlier than PMF” and “concentrate on the system not the mannequin,” to assist groups work out the place to allocate scarce assets. We additionally recommend a roadmap for iterating towards a fantastic product. This last set of classes solutions the next questions:

  1. Constructing vs. Shopping for: When must you prepare your individual fashions, and when must you leverage current APIs? The reply is, as all the time, “it relies upon.” We share what it relies on.
  2. Iterating to One thing Nice: How will you create a long-lasting aggressive edge that goes past simply utilizing the newest fashions? We talk about the significance of constructing a sturdy system across the mannequin and specializing in delivering memorable, sticky experiences.
  3. Human-Centered AI: How will you successfully combine LLMs into human workflows to maximise productiveness and happiness? We emphasize the significance of constructing AI instruments that assist and improve human capabilities fairly than trying to switch them totally.
  4. Getting Began: What are the important steps for groups embarking on constructing an LLM product? We define a fundamental playbook that begins with immediate engineering, evaluations, and information assortment.
  5. The Way forward for Low-Price Cognition: How will the quickly reducing prices and rising capabilities of LLMs form the way forward for AI purposes? We study historic tendencies and stroll via a easy technique to estimate when sure purposes would possibly turn out to be economically possible.
  6. From Demos to Merchandise: What does it take to go from a compelling demo to a dependable, scalable product? We emphasize the necessity for rigorous engineering, testing, and refinement to bridge the hole between prototype and manufacturing.

To reply these tough questions, let’s suppose step-by-step…

Technique: Constructing with LLMs with out Getting Out-Maneuvered

Profitable merchandise require considerate planning and hard prioritization, not countless prototyping or following the newest mannequin releases or tendencies. On this last part, we glance across the corners and take into consideration the strategic concerns for constructing nice AI merchandise. We additionally study key trade-offs groups will face, like when to construct and when to purchase, and recommend a “playbook” for early LLM software growth technique.

No GPUs earlier than PMF

To be nice, your product must be greater than only a skinny wrapper round any individual else’s API. However errors in the other way may be much more pricey. The previous 12 months has additionally seen a mint of enterprise capital, together with an eye-watering six-billion-dollar Sequence A, spent on coaching and customizing fashions with out a clear product imaginative and prescient or goal market. On this part, we’ll clarify why leaping instantly to coaching your individual fashions is a mistake and think about the position of self-hosting.

Coaching from scratch (nearly) by no means is sensible

For many organizations, pretraining an LLM from scratch is an impractical distraction from constructing merchandise.

As thrilling as it’s and as a lot because it looks as if everybody else is doing it, creating and sustaining machine studying infrastructure takes loads of assets. This contains gathering information, coaching and evaluating fashions, and deploying them. If you happen to’re nonetheless validating product-market match, these efforts will divert assets from creating your core product. Even in the event you had the compute, information, and technical chops, the pretrained LLM might turn out to be out of date in months.

Think about the case of BloombergGPT, an LLM particularly educated for monetary duties. The mannequin was pretrained on 363B tokens and required a heroic effort by 9 full-time workers, 4 from AI Engineering and 5 from ML Product and Analysis. Regardless of this effort, it was outclassed by gpt-3.5-turbo and gpt-4 on these monetary duties inside a 12 months.

This story and others prefer it means that for many sensible purposes, pretraining an LLM from scratch, even on domain-specific information, just isn’t the very best use of assets. As a substitute, groups are higher off fine-tuning the strongest open supply fashions accessible for his or her particular wants.

There are after all exceptions. One shining instance is Replit’s code mannequin, educated particularly for code-generation and understanding. With pretraining, Replit was capable of outperform different fashions of huge sizes similar to CodeLlama7b. However as different, more and more succesful fashions have been launched, sustaining utility has required continued funding.

Don’t fine-tune till you’ve confirmed it’s mandatory

For many organizations, fine-tuning is pushed extra by FOMO than by clear strategic pondering.

Organizations put money into fine-tuning too early, attempting to beat the “simply one other wrapper” allegations. In actuality, fine-tuning is heavy equipment, to be deployed solely after you’ve collected loads of examples that persuade you different approaches gained’t suffice.

A 12 months in the past, many groups have been telling us they have been excited to fine-tune. Few have discovered product-market match and most remorse their choice. If you happen to’re going to fine-tune, you’d higher be actually assured that you simply’re set as much as do it time and again as base fashions enhance—see the “The mannequin isn’t the product” and “Construct LLMOps” beneath.

When would possibly fine-tuning truly be the appropriate name? If the use case requires information not accessible within the principally open web-scale datasets used to coach current fashions—and in the event you’ve already constructed an MVP that demonstrates the prevailing fashions are inadequate. However watch out: if nice coaching information isn’t available to the mannequin builders, the place are you getting it?

In the end, keep in mind that LLM-powered purposes aren’t a science honest challenge; funding in them needs to be commensurate with their contribution to your online business’ strategic targets and its aggressive differentiation.

Begin with inference APIs, however don’t be afraid of self-hosting

With LLM APIs, it’s simpler than ever for startups to undertake and combine language modeling capabilities with out coaching their very own fashions from scratch. Suppliers like Anthropic and OpenAI supply normal APIs that may sprinkle intelligence into your product with just some traces of code. Through the use of these providers, you possibly can scale back the trouble spent and as a substitute concentrate on creating worth to your prospects—this lets you validate concepts and iterate towards product-market match sooner.

However, as with databases, managed providers aren’t the appropriate match for each use case, particularly as scale and necessities improve. Certainly, self-hosting would be the solely manner to make use of fashions with out sending confidential/personal information out of your community, as required in regulated industries like healthcare and finance or by contractual obligations or confidentiality necessities.

Moreover, self-hosting circumvents limitations imposed by inference suppliers, like price limits, mannequin deprecations, and utilization restrictions. As well as, self-hosting provides you full management over the mannequin, making it simpler to assemble a differentiated, high-quality system round it. Lastly, self-hosting, particularly of fine-tunes, can scale back price at giant scale. For instance, BuzzFeed shared how they fine-tuned open supply LLMs to scale back prices by 80%.

Iterate to one thing nice

To maintain a aggressive edge in the long term, it’s essential suppose past fashions and think about what is going to set your product aside. Whereas pace of execution issues, it shouldn’t be your solely benefit.

The mannequin isn’t the product; the system round it’s

For groups that aren’t constructing fashions, the fast tempo of innovation is a boon as they migrate from one SOTA mannequin to the subsequent, chasing beneficial properties in context measurement, reasoning functionality, and price-to-value to construct higher and higher merchandise.

This progress is as thrilling as it’s predictable. Taken collectively, this implies fashions are more likely to be the least sturdy element within the system.

As a substitute, focus your efforts on what’s going to supply lasting worth, similar to:

  • Analysis chassis: To reliably measure efficiency in your process throughout fashions
  • Guardrails: To forestall undesired outputs irrespective of the mannequin
  • Caching: To cut back latency and price by avoiding the mannequin altogether
  • Information flywheel: To energy the iterative enchancment of every thing above

These elements create a thicker moat of product high quality than uncooked mannequin capabilities.

However that doesn’t imply constructing on the software layer is danger free. Don’t level your shears on the identical yaks that OpenAI or different mannequin suppliers might want to shave in the event that they need to present viable enterprise software program.

For instance, some groups invested in constructing customized tooling to validate structured output from proprietary fashions; minimal funding right here is necessary, however a deep one just isn’t a superb use of time. OpenAI wants to make sure that once you ask for a operate name, you get a sound operate name—as a result of all of their prospects need this. Make use of some “strategic procrastination” right here, construct what you completely want and await the apparent expansions to capabilities from suppliers.

Construct belief by beginning small

Constructing a product that tries to be every thing to everyone seems to be a recipe for mediocrity. To create compelling merchandise, firms have to focus on constructing memorable, sticky experiences that maintain customers coming again.

Think about a generic RAG system that goals to reply any query a consumer would possibly ask. The shortage of specialization signifies that the system can’t prioritize latest info, parse domain-specific codecs, or perceive the nuances of particular duties. Consequently, customers are left with a shallow, unreliable expertise that doesn’t meet their wants.

To handle this, concentrate on particular domains and use instances. Slender the scope by going deep fairly than broad. This can create domain-specific instruments that resonate with customers. Specialization additionally means that you can be upfront about your system’s capabilities and limitations. Being clear about what your system can and can’t do demonstrates self-awareness, helps customers perceive the place it might add essentially the most worth, and thus builds belief and confidence within the output.

Construct LLMOps, however construct it for the appropriate motive: sooner iteration

DevOps just isn’t basically about reproducible workflows or shifting left or empowering two pizza groups—and it’s undoubtedly not about writing YAML information.

DevOps is about shortening the suggestions cycles between work and its outcomes in order that enhancements accumulate as a substitute of errors. Its roots return, by way of the Lean Startup motion, to Lean manufacturing and the Toyota Manufacturing System, with its emphasis on Single Minute Trade of Die and Kaizen.

MLOps has tailored the type of DevOps to ML. We’ve got reproducible experiments and we’ve all-in-one suites that empower mannequin builders to ship. And Lordy, do we’ve YAML information.

However as an trade, MLOps didn’t adapt the operate of DevOps. It didn’t shorten the suggestions hole between fashions and their inferences and interactions in manufacturing.

Hearteningly, the sphere of LLMOps has shifted away from fascinated with hobgoblins of little minds like immediate administration and towards the laborious issues that block iteration: manufacturing monitoring and continuous enchancment, linked by analysis.

Already, we’ve interactive arenas for impartial, crowd-sourced analysis of chat and coding fashions—an outer loop of collective, iterative enchancment. Instruments like LangSmith, Log10, LangFuse, W&B Weave, HoneyHive, and extra promise to not solely acquire and collate information about system outcomes in manufacturing but in addition to leverage them to enhance these techniques by integrating deeply with growth. Embrace these instruments or construct your individual.

Don’t construct LLM options you should buy

Most profitable companies usually are not LLM companies. Concurrently, most companies have alternatives to be improved by LLMs.

This pair of observations typically misleads leaders into swiftly retrofitting techniques with LLMs at elevated price and decreased high quality and releasing them as ersatz, vainness “AI” options, full with the now-dreaded sparkle icon. There’s a greater manner: concentrate on LLM purposes that actually align along with your product targets and improve your core operations.

Think about just a few misguided ventures that waste your crew’s time:

  • Constructing customized text-to-SQL capabilities for your online business
  • Constructing a chatbot to speak to your documentation
  • Integrating your organization’s data base along with your buyer assist chatbot

Whereas the above are the hellos-world of LLM purposes, none of them make sense for nearly any product firm to construct themselves. These are normal issues for a lot of companies with a big hole between promising demo and reliable element—the customary area of software program firms. Investing useful R&D assets on normal issues being tackled en masse by the present Y Combinator batch is a waste.

If this appears like trite enterprise recommendation, it’s as a result of within the frothy pleasure of the present hype wave, it’s simple to mistake something “LLM” as cutting-edge accretive differentiation, lacking which purposes are already outdated hat.

AI within the loop; people on the middle

Proper now, LLM-powered purposes are brittle. They required an unbelievable quantity of safe-guarding and defensive engineering and stay laborious to foretell. Moreover, when tightly scoped, these purposes may be wildly helpful. Which means that LLMs make glorious instruments to speed up consumer workflows.

Whereas it could be tempting to think about LLM-based purposes absolutely changing a workflow or standing in for a job operate, at this time the best paradigm is a human-computer centaur (c.f. Centaur chess). When succesful people are paired with LLM capabilities tuned for his or her fast utilization, productiveness and happiness doing duties may be massively elevated. One of many flagship purposes of LLMs, GitHub Copilot, demonstrated the ability of those workflows:

“Total, builders advised us they felt extra assured as a result of coding is less complicated, extra error-free, extra readable, extra reusable, extra concise, extra maintainable, and extra resilient with GitHub Copilot and GitHub Copilot Chat than after they’re coding with out it.”
Mario Rodriguez, GitHub

For individuals who have labored in ML for a very long time, it’s possible you’ll leap to the thought of “human-in-the-loop,” however not so quick: HITL machine studying is a paradigm constructed on human consultants guaranteeing that ML fashions behave as predicted. Whereas associated, right here we’re proposing one thing extra refined. LLM pushed techniques shouldn’t be the first drivers of most workflows at this time; they need to merely be a useful resource.

By centering people and asking how an LLM can assist their workflow, this results in considerably totally different product and design choices. In the end, it’s going to drive you to construct totally different merchandise than opponents who attempt to quickly offshore all accountability to LLMs—higher, extra helpful, and fewer dangerous merchandise.

Begin with prompting, evals, and information assortment

The earlier sections have delivered a hearth hose of strategies and recommendation. It’s lots to soak up. Let’s think about the minimal helpful set of recommendation: if a crew desires to construct an LLM product, the place ought to they start?

During the last 12 months, we’ve seen sufficient examples to begin turning into assured that profitable LLM purposes comply with a constant trajectory. We stroll via this fundamental “getting began” playbook on this part. The core concept is to begin easy and solely add complexity as wanted. A good rule of thumb is that every stage of sophistication usually requires no less than an order of magnitude extra effort than the one earlier than it. With this in thoughts…

Immediate engineering comes first

Begin with immediate engineering. Use all of the strategies we mentioned within the techniques part earlier than. Chain-of-thought, n-shot examples, and structured enter and output are nearly all the time a good suggestion. Prototype with essentially the most extremely succesful fashions earlier than attempting to squeeze efficiency out of weaker fashions.

Provided that immediate engineering can not obtain the specified stage of efficiency must you think about fine-tuning. This can come up extra typically if there are nonfunctional necessities (e.g., information privateness, full management, and price) that block the usage of proprietary fashions and thus require you to self-host. Simply ensure those self same privateness necessities don’t block you from utilizing consumer information for fine-tuning!

Construct evals and kickstart an information flywheel

Even groups which might be simply getting began want evals. In any other case, you gained’t know whether or not your immediate engineering is adequate or when your fine-tuned mannequin is able to exchange the bottom mannequin.

Efficient evals are particular to your duties and mirror the meant use instances. The primary stage of evals that we suggest is unit testing. These easy assertions detect recognized or hypothesized failure modes and assist drive early design choices. Additionally see different task-specific evals for classification, summarization, and many others.

Whereas unit checks and model-based evaluations are helpful, they don’t exchange the necessity for human analysis. Have individuals use your mannequin/product and supply suggestions. This serves the twin function of measuring real-world efficiency and defect charges whereas additionally gathering high-quality annotated information that can be utilized to fine-tune future fashions. This creates a constructive suggestions loop, or information flywheel, which compounds over time:

  • Use human analysis to evaluate mannequin efficiency and/or discover defects
  • Use the annotated information to fine-tune the mannequin or replace the immediate

For instance, when auditing LLM-generated summaries for defects we would label every sentence with fine-grained suggestions figuring out factual inconsistency, irrelevance, or poor fashion. We will then use these factual inconsistency annotations to prepare a hallucination classifier or use the relevance annotations to coach a reward mannequin to attain on relevance. As one other instance, LinkedIn shared about its success with utilizing model-based evaluators to estimate hallucinations, accountable AI violations, coherence, and many others. in its write-up.

By creating belongings that compound their worth over time, we improve constructing evals from a purely operational expense to a strategic funding and construct our information flywheel within the course of.

The high-level development of low-cost cognition

In 1971, the researchers at Xerox PARC predicted the longer term: the world of networked private computer systems that we at the moment are residing in. They helped delivery that future by enjoying pivotal roles within the invention of the applied sciences that made it attainable, from Ethernet and graphics rendering to the mouse and the window.

However additionally they engaged in a easy train: they checked out purposes that have been very helpful (e.g., video shows) however weren’t but economical (i.e., sufficient RAM to drive a video show was many 1000’s of {dollars}). Then they checked out historic worth tendencies for that expertise (à la Moore’s legislation) and predicted when these applied sciences would turn out to be economical.

We will do the identical for LLM applied sciences, although we don’t have one thing fairly as clear as transistors-per-dollar to work with. Take a preferred, long-standing benchmark, just like the Massively-Multitask Language Understanding dataset, and a constant enter method (five-shot prompting). Then, evaluate the associated fee to run language fashions with numerous efficiency ranges on this benchmark over time.

For a set price, capabilities are quickly rising. For a set functionality stage, prices are quickly reducing. Created by coauthor Charles Frye utilizing public information on Might 13, 2024.

Within the 4 years for the reason that launch of OpenAI’s davinci mannequin as an API, the associated fee for operating a mannequin with equal efficiency on that process on the scale of 1 million tokens (about 100 copies of this doc) has dropped from $20 to lower than 10¢—a halving time of simply six months. Equally, the associated fee to run Meta’s LLama 3 8B by way of an API supplier or by yourself is simply 20¢ per million tokens as of Might 2024, and it has related efficiency to OpenAI’s text-davinci-003, the mannequin that enabled ChatGPT to shock the world. That mannequin additionally price about $20 per million tokens when it was launched in late November 2023. That’s two orders of magnitude in simply 18 months—the identical timeframe during which Moore’s legislation predicts a mere doubling.

Now, let’s think about an software of LLMs that may be very helpful (powering generative online game characters, à la Park et al.) however just isn’t but economical. (Their price was estimated at $625 per hour right here.) Since that paper was revealed in August 2023, the associated fee has dropped roughly one order of magnitude, to $62.50 per hour. We’d count on it to drop to $6.25 per hour in one other 9 months.

In the meantime, when Pac-Man was launched in 1980, $1 of at this time’s cash would purchase you a credit score, good to play for a couple of minutes or tens of minutes—name it six video games per hour, or $6 per hour. This serviette math suggests {that a} compelling LLM-enhanced gaming expertise will turn out to be economical a while in 2025.

These tendencies are new, only some years outdated. However there may be little motive to count on this course of to decelerate within the subsequent few years. Whilst we maybe expend low-hanging fruit in algorithms and datasets, like scaling previous the “Chinchilla ratio” of ~20 tokens per parameter, deeper improvements and investments inside the information middle and on the silicon layer promise to choose up slack.

And that is maybe a very powerful strategic truth: what’s a very infeasible ground demo or analysis paper at this time will turn out to be a premium function in just a few years after which a commodity shortly after. We must always construct our techniques, and our organizations, with this in thoughts.

Sufficient 0 to 1 Demos, It’s Time for 1 to N Merchandise

We get it; constructing LLM demos is a ton of enjoyable. With just some traces of code, a vector database, and a rigorously crafted immediate, we create ✨magic ✨. And prior to now 12 months, this magic has been in comparison with the web, the smartphone, and even the printing press.

Sadly, as anybody who has labored on delivery real-world software program is aware of, there’s a world of distinction between a demo that works in a managed setting and a product that operates reliably at scale.

Take, for instance, self-driving automobiles. The primary automobile was pushed by a neural community in 1988. Twenty-five years later, Andrej Karpathy took his first demo experience in a Waymo. A decade after that, the corporate acquired its driverless allow. That’s thirty-five years of rigorous engineering, testing, refinement, and regulatory navigation to go from prototype to industrial product.

Throughout totally different elements of trade and academia, we’ve keenly noticed the ups and downs for the previous 12 months: 12 months 1 of N for LLM purposes. We hope that the teachings we’ve discovered—from techniques like rigorous operational strategies for constructing groups to strategic views like which capabilities to construct internally—aid you in 12 months 2 and past, as all of us construct on this thrilling new expertise collectively.

In regards to the authors

Eugene Yan designs, builds, and operates machine studying techniques that serve prospects at scale. He’s presently a Senior Utilized Scientist at Amazon the place he builds RecSys for thousands and thousands worldwide and applies LLMs to serve prospects higher. Beforehand, he led machine studying at Lazada (acquired by Alibaba) and a Healthtech Sequence A. He writes & speaks about ML, RecSys, LLMs, and engineering at eugeneyan.com and ApplyingML.com.

Bryan Bischof is the Head of AI at Hex, the place he leads the crew of engineers constructing Magic – the information science and analytics copilot. Bryan has labored all around the information stack main groups in analytics, machine studying engineering, information platform engineering, and AI engineering. He began the information crew at Blue Bottle Espresso, led a number of initiatives at Sew Repair, and constructed the information groups at Weights and Biases. Bryan beforehand co-authored the e-book Constructing Manufacturing Advice Programs with O’Reilly, and teaches Information Science and Analytics within the graduate faculty at Rutgers. His Ph.D. is in pure arithmetic.

Charles Frye teaches individuals to construct AI purposes. After publishing analysis in psychopharmacology and neurobiology, he acquired his Ph.D. on the College of California, Berkeley, for dissertation work on neural community optimization. He has taught 1000’s all the stack of AI software growth, from linear algebra fundamentals to GPU arcana and constructing defensible companies, via instructional and consulting work at Weights and Biases, Full Stack Deep Studying, and Modal.

Hamel Husain is a machine studying engineer with over 25 years of expertise. He has labored with progressive firms similar to Airbnb and GitHub, which included early LLM analysis utilized by OpenAI for code understanding. He has additionally led and contributed to quite a few fashionable open-source machine-learning instruments. Hamel is presently an impartial guide serving to firms operationalize Massive Language Fashions (LLMs) to speed up their AI product journey.

Jason Liu is a distinguished machine studying guide recognized for main groups to efficiently ship AI merchandise. Jason’s technical experience covers personalization algorithms, search optimization, artificial information era, and MLOps techniques.

His expertise contains firms like Sew Repair, the place he created a suggestion framework and observability instruments that dealt with 350 million every day requests. Further roles have included Meta, NYU, and startups similar to Limitless AI and Trunk Instruments.

Shreya Shankar is an ML engineer and PhD pupil in laptop science at UC Berkeley. She was the primary ML engineer at 2 startups, constructing AI-powered merchandise from scratch that serve 1000’s of customers every day. As a researcher, her work focuses on addressing information challenges in manufacturing ML techniques via a human-centered method. Her work has appeared in high information administration and human-computer interplay venues like VLDB, SIGMOD, CIDR, and CSCW.

Contact Us

We might love to listen to your ideas on this publish. You may contact us at contact@applied-llms.org. Many people are open to numerous types of consulting and advisory. We’ll route you to the proper professional(s) upon contact with us if applicable.

Acknowledgements

This sequence began as a dialog in a bunch chat, the place Bryan quipped that he was impressed to put in writing “A 12 months of AI Engineering”. Then, ✨magic✨ occurred within the group chat (see picture beneath), and we have been all impressed to chip in and share what we’ve discovered to this point.

The authors wish to thank Eugene for main the majority of the doc integration and general construction along with a big proportion of the teachings. Moreover, for main enhancing tasks and doc path. The authors wish to thank Bryan for the spark that led to this writeup, restructuring the write-up into tactical, operational, and strategic sections and their intros, and for pushing us to suppose larger on how we might attain and assist the neighborhood. The authors wish to thank Charles for his deep dives on price and LLMOps, in addition to weaving the teachings to make them extra coherent and tighter—you could have him to thank for this being 30 as a substitute of 40 pages! The authors recognize Hamel and Jason for his or her insights from advising shoppers and being on the entrance traces, for his or her broad generalizable learnings from shoppers, and for deep data of instruments. And at last, thanks Shreya for reminding us of the significance of evals and rigorous manufacturing practices and for bringing her analysis and authentic outcomes to this piece.

Lastly, the authors wish to thank all of the groups who so generously shared your challenges and classes in your individual write-ups which we’ve referenced all through this sequence, together with the AI communities to your vibrant participation and engagement with this group.



Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

[td_block_social_counter facebook="tagdiv" twitter="tagdivofficial" youtube="tagdiv" style="style8 td-social-boxed td-social-font-icons" tdc_css="eyJhbGwiOnsibWFyZ2luLWJvdHRvbSI6IjM4IiwiZGlzcGxheSI6IiJ9LCJwb3J0cmFpdCI6eyJtYXJnaW4tYm90dG9tIjoiMzAiLCJkaXNwbGF5IjoiIn0sInBvcnRyYWl0X21heF93aWR0aCI6MTAxOCwicG9ydHJhaXRfbWluX3dpZHRoIjo3Njh9" custom_title="Stay Connected" block_template_id="td_block_template_8" f_header_font_family="712" f_header_font_transform="uppercase" f_header_font_weight="500" f_header_font_size="17" border_color="#dd3333"]
- Advertisement -spot_img

Latest Articles