

TL;DR:
Industry benchmarks for ecommerce are hard to come by. Most of what's out there is self-reported, survey-based, or too aggregated to be usable. Teams are left wondering whether their AI adoption is on par with industry standards or if their response times are costing them revenue.
That's a gap we're in a unique position to close.
Gorgias processes millions of customer conversations across thousands of ecommerce brands every day. This has given us a rare, unfiltered view into how the industry operates. But until now, we’ve kept those insights largely internal.
Today, we're making it public with the Ecom Lab.
The result is years of first-party data from thousands of ecommerce brands, packaged into findings that give teams a real foundation to build their strategy on.
The Ecom Lab is Gorgias's public research hub for ecommerce. It publishes insights and reports on AI adoption, support performance, financial impact, and industry trends.
The goal is simple: give teams a real baseline to measure against and to uncover the industry's inner workings.
Metrics that actually move decisions.
The Ecom Lab publishes metrics that matter to ecommerce professionals, including AI adoption rates, first response times, CSAT scores, conversion rates, and ticket intents, all broken down by brand size, GMV tier, and industry vertical.
For the first time, teams can see exactly where they stand in comparison to the broader market.
AI is Everywhere reveals why roughly 4 in 5 ecommerce brands still haven't deployed AI in customer-facing support.
Stop Benchmarking Against the Average argues that support teams should benchmark response times against their specific industry vertical rather than the overall average.
Most Brands are Overpaying for Support breaks down the actual cost of support ticket volume and what happens when AI handles the load.
Four months ago, our analysts were dealing with a barrage of questions. "What's our ARR by segment?" "Build me a dashboard for this quarter's pipeline." Quick asks piled up behind complex deep dives. Stakeholders waited for answers that should have taken seconds, and analysts spent their time fielding requests instead of doing the strategic work that creates the most value.
Today, anyone at Gorgias can ask a question in plain language and get an accurate, contextualized response in seconds. Not from a colleague or dashboard, nor from a generic answer from the internet. But a response built on our business context. We call it Cortex, our flagship internal AI agent.
In two months, Cortex went from an idea to fielding thousands of questions every week, recommending actions across the business, and deprecating the need for manual dashboard creation. While most companies right now are treating AI as an initiative — at Gorgias, AI is already part of how we work. 72% of Gorgias employees use Cortex each week, and that number is only growing.
We didn’t achieve this by simply plugging a large language model into our stack. LLMs are a critical part of the equation, but they aren't the driving force — it’s everything else under the hood: the infrastructure, context, platform architecture, and the team that brings it all together.

The instinct across many companies today is to start with the model, pick a provider to solve a specific challenge, or invest heavily in getting the data right. All reasonable starting points, but most of them solve for one use case. Underneath that approach is a framing problem: seeing AI as an initiative — something you assign and measure. Seeing AI as another tool your company uses versus how your company operates.
We started somewhere different. Every company is built on four pillars: customers, people, product, and decisions. AI investments tend to place heavy emphasis on the first three. We started with the fourth. Our bet was that if we built everything around the need to make effective decisions first, asking what Gorgias needed to know to operate well, then our AI would become dramatically more powerful.
Cortex is our flagship internal AI agent, and the product where we established the tenets that now run through everything else we build: composable and modular infrastructure, governed context, and accessible from wherever decisions happen. Cortex lives in Slack, as well as across LLM vendors, in its own browser extension, and even on its own dedicated internal site.
Cortex doesn’t stop at answering questions. It can read and write to Notion, file Linear tasks, create HTML apps, automate signal delivery, and more. It operates across every layer of our stack, from dashboards to data pipelines, because we designed it as one integrated system. It is this connection that adds remarkable depth to what people can ask, and what they get in return.

A Sales Lead is pitching and asks Cortex for the full picture of the merchant. In a customized PDF, Cortex lists coverage gaps, pre-sale intent signals, and product fit options. Everything the sales lead needs to walk in with confidence.
A Senior Product leader asks, "How are we performing against OKR #1, and what can my team do to help accelerate it?" Cortex returns a full ARR breakdown, projected end-of-month attainment, segment-level findings, and connects it all back to company-level strategies. A suite of recommendations customized to the leader, the performance, and the signals that bridge how they can support our goals. The kind of answer that used to take someone a week to put together.
These aren't simple lookup queries. They require deep business context spanning multiple areas. Cortex handles these because its Decision Engine gives it the information to reason against governed data, metric definitions, and business context, turning a generic answer into a credible one.
Overnight, teams have built Cortex into how they work. They’re spending less time searching and more time finding answers, not because they were told to, but because Cortex reduced the distance between question and decision.
Cortex’s modular infrastructure allows us to experiment and add new capabilities freely. We’ve already built two more internal AI agents made for entirely different use cases, but using the same Decision Engine as Cortex.
GAIA, our internal experimentation AI Agent, helps our customers identify opportunities in their AI Agent Guidance design. It takes institutional knowledge across our teams and turns it into a scalable system that drives automation and value to our customers. Our CEO, Romain Lapeyre, has been its most vocal advocate since day one.
When we needed a platform for investor readiness and board preparation, we built Oracle. Our board decks and talk tracks are informed and built with the same AI, and our numbers are validated every step of the way.
We’re continuing to expand new AI agents internally, exploring how they can create value for customers and our own teams.
When AI handles thousands of analytical questions each week, the highest-value work for a data team shifts permanently. Late 2025, we repositioned from a Data Analytics function into a Decision Intelligence function — a structural change in what we own and how we operate.
Today, our analysts focus on the most sensitive, complex, and forward-looking decisions and analyses. They partner more deeply with stakeholders by driving next steps from signals. They're even building entirely new capabilities that didn't exist in their role descriptions months ago. Things like AI skills for Cortex, context curation, and insight and recommendation delivery. The role of the analyst hasn't diminished. It's expanded to encompass the most meaningful work an analyst can do: driving outcomes and ensuring those decisions can achieve them.

Our business support model has changed, too. Instead of embedding analysts and dedicated engineers within functional teams, we align capacity to the highest-impact company objectives and move fluidly across them. This model works even better because Decision Intelligence brings together both analytics and engineering teams under one roof.
Elliot Trabac leads our Data, Context and AI Engineering teams. The Decision Engine, Cortex, GAIA, and the platforms I've described exist because of the infrastructure his team innovated and built from the ground up. Noemie Happi Nono leads our Decision Strategy and Operations team, driving decision outcomes with stakeholders, advancing the development of Cortex skills and capabilities, and pushing into new areas of analysis every day.
Together, they're shaping what a modern data function looks like when AI becomes a standard building block for how a company operates.
The question of ROI is long gone. AI has opened the floodgates to more trusted and meaningful signals than ever. The natural next evolution is Proactive Intelligence, signals surfaced toward what you need to know, before you ask. And we're already building this because our architecture is designed to support it.
In the coming weeks, members of the Decision Intelligence team will go deeper into themes I've touched on here. Yochan Khoi, a Senior Analytics Engineer on our team, recently published a technical walkthrough of our context layer and will go further into building context strategies that scale. Others will cover infrastructure, analytical partnerships, evolving data assets into decision assets, and the cost and efficiency gains that make sustained AI investment viable.
AI hasn't changed the most important element of data and analytics functions — delivering outcomes — but it has raised the bar for what it looks like and how far we can take it. We’re just getting started.
The best in CX and ecommerce, right to your inbox

TL;DR:
The page-based shopping experience dominated for decades. Customers would search, browse, compare, abandon, get retargeted, return, and eventually buy (sometimes).
That journey is no longer the only option.
Shoppers are turning to chat, messaging, and AI-powered tools to find what they need. Instead of clicking through product pages or reading static FAQs, they ask questions, have back-and-forth conversations, and get answers that move them closer to a purchase in real time. The path to checkout has changed, and the brands that recognize this are pulling ahead.
Read our 2026 State of Conversational Commerce Report to learn more about conversation commerce trends from 400 ecommerce decision-makers and 16,000+ ecommerce brands using Gorgias.
{{lead-magnet-1}}
The traditional shopping journey was a solo experience. A shopper had a need, searched for options, browsed across sessions, and eventually made a decision — often days later, after being retargeted multiple times. Support only entered the picture after the purchase.

The conversation-led journey collapses that timeline:
What used to take days now takes minutes. Discovery, evaluation, and purchase happen in a single thread.
79% of brands agree that AI-driven conversational commerce has increased sales and purchase rates in their business. When brands were asked to rank the highest-return areas:
Those numbers reflect something important: the value of conversation compounds. Faster support reduces friction. Better retention raises lifetime value. More confident shoppers buy more often and spend more per order.
The brands seeing the biggest returns aren't just using AI to deflect tickets. They're using it to create one-to-one shopping experiences at scale.
Looking at AI-only influenced orders across key verticals like Apparel and Accessories, Food and Beverages, Health and Beauty, Home and Garden, and Sporting Goods, the growth across a single year was significant.





Across industries, ecommerce brands saw AI step into conversations, reduce shopper hesitation, and drive higher QoQ conversion rates.
Learn more about AI-powered revenue generation in the full 2026 Conversational Commerce Report.
84% of brands say the strategic importance of conversational commerce is higher than it was a year ago. 82% agree it will be mainstream in their sector within two years.

That shift is registering at the leadership level because of what conversational commerce does to the buying experience. Creating one-to-one touchpoints earlier in the journey drives higher AOV, shorter buying cycles, and stronger purchase rates. Shoppers who get real-time answers to their questions are more confident.
TUSHY, known for eco-friendly bidets and bathroom essentials, is a useful example of what happens when you take conversational commerce seriously.
Bidets aren't an impulse purchase. Shoppers have real questions about fit, compatibility, and installation. Those questions used to go unanswered until the CX team could respond, often after the customer had abandoned the cart.
TUSHY used Gorgias's AI Agent and shopping assistant capabilities to automate pre-sales support. AI Agent engaged shoppers in real-time conversations, addressed their concerns directly, and built confidence at the moment of highest intent.
This resulted in a 190% increase in chat-based purchases, a 13x return on investment, and twice the purchase rate of human agents.
You don't need to overhaul your entire operation to start seeing results. The most effective approach is to start where the impact is clearest and expand from there.
A few places to begin:
Want to see the full picture of where conversational commerce is headed in 2026? Read the full report to explore the data, trends, and strategies shaping the next era of ecommerce.
{{lead-magnet-1}}

TL;DR:
A year ago, ecommerce brands were still debating whether AI was worth the investment. That debate is over. Today, nearly every ecommerce professional uses AI to do their job.
The shift isn't just about adoption. It's about what AI is used for and how brands measure its impact. Support automation was the entry point. Now, AI is embedded across the full operation, from product recommendations to inventory control to real-time shopping conversations.
In our 2026 State of Conversational Commerce Report, we break down trends on AI usage among 400 ecommerce decision-makers and 16,000+ ecommerce brands using Gorgias.
{{lead-magnet-1}}
If we rewind 12 months ago, the industry was still split on AI. Some ecommerce professionals were excited, but most were still hesitant. In 2024, 69% of ecommerce professionals used AI in their roles. By 2025, that number reached 77%. In 2026, it hit 96%.

The confidence numbers back it up. 71% of brands say they are confident using AI for ecommerce, and 73% are satisfied with its business impact.
In early 2025, only 30% of ecommerce professionals rated their excitement for AI at 10/10. Today, zero percent of respondents describe themselves as hesitant about AI.

Using AI in ecommerce is not new. In fact, it dates back to the 1980s with the invention of algorithms and expert systems. And if you’ve ever leveraged similar product recommendations or chatbots, you’ve already integrated AI into your ecommerce stack.
Modern AI is far more sophisticated.
With the rise of agentic commerce and conversational AI, brands began leveraging AI agents to automate the processing of repetitive support tickets. That’s still happening today, but the scope has expanded beyond the support queue.

Ecommerce brands are deploying AI across every layer of their operation:
When brands were asked which channels contribute most to their AI success, conversational channels dominated. Social media messaging led at 78%, followed by SMS at 70%, and website live chat at 51%. Shoppers want fast, personal conversations, and AI is the best way to deliver that at scale.
Learn more about AI adoption, perception, and use case trends in the full 2026 Conversational Commerce Report.
For decades, customer support success meant fast response times and high satisfaction scores. Those are still important indicators of success, but leading brands are adding revenue-focused metrics to their dashboards.
91% of brands still track CSAT as a measure of AI's impact. But 60% now include AOV as a top indicator, and higher-revenue brands earning $20M+ are focusing on metrics like total operating expenses, cost per resolution, incremental revenue, and one-touch ticket rate.

AI can now start a conversation, ease customer doubts, sell, upsell, and recover abandoned carts in a single conversation. When you’re only measuring CSAT, you’re ignoring the real ROI of conversational AI investment.
Virtual shopping assistants now proactively engage shoppers, adapt to their needs in real time, and offer contextual product recommendations and upsells. When the moment calls for it, they can close the deal with a targeted discount.
Gorgias brands using AI Agent's shopping assistant capabilities nearly doubled their purchase rates and converted 20–50% better than those using AI Agent for support only.
Orthofeet, the largest provider of orthopedic footwear in the US, is a concrete example of this in practice. Using Gorgias, they achieved:
The data tells a clear story: AI has evolved beyond a tool for handling tier 1 support tickets. It’s a core part of your revenue generation strategy.
57% of brands are already using AI for 26–50% of all customer interactions, and 37% expect that share to rise to 51–75% within the next two years. The brands building toward that range now are the ones who will have the operational advantage when it matters most.
The practical question isn't whether to invest in AI. It's where to focus first. Based on where brands are seeing the most impact, three priorities stand out:
Want to go deeper on the full 2026 conversational commerce trends? Read the complete report for data across every major AI use case in ecommerce.
{{lead-magnet-1}}

TL;DR:
The way shoppers buy online has shifted and customers are at the center.
They no longer want to scroll through product pages, dig through FAQs, or wait 24 hours for an email reply. They open a conversation, ask a specific question, and expect a useful answer in seconds. Brands that can’t deliver these experiences at scale are seeing customer hesitation turn into abandoned carts and lost revenue.
This shift has a name: conversational commerce. It's the practice of using real-time, two-way conversations as your primary sales channel, through chat, AI agents, messaging apps, and voice.
What started as an experiment for early adopters has become a key growth lever, with 84% of ecommerce brands treating conversational commerce as a strategic pillar this year vs. last year.

We surveyed 400 ecommerce decision-makers across North America, the U.K., and Europe to understand how conversational commerce and AI are reshaping the ecommerce landscape. These findings are complemented by aggregated and anonymized internal Gorgias platform data from 16,000+ ecommerce brands.
The State of Conversational Commerce in 2026 trends report breaks down all of the findings, including five key trends shaping the ecommerce landscape.
{{lead-magnet-1}}
A few years ago, adding an AI chatbot to your site that could provide tracking links and Help Center article recommendations was a differentiator. Today, it's table stakes. McKinsey found that 71% of shoppers expect personalized experiences, and 76% get frustrated when they don't get them.
Right now, most ecommerce professionals use AI, with 93% having used it for at least 1 year. Enthusiasm is accelerating quickly, with only 30% of ecommerce professionals rating their excitement for AI at 10/10 in April 2025. Similarly, while AI adoption rose steadily year over year, it reached a clear peak in 2026.

The use cases driving this adoption are practical and high-volume:

These are the tickets that flood brands’ inboxes every day. AI agents resolve them instantly, without pulling teams away from conversations that actually require human judgment.
Explore AI adoption and use case data in more depth in the full report.
The traditional ecommerce funnel, visit site, browse products, add to cart, check out, is losing ground. Shoppers now discover products on Instagram, ask questions via direct message, and complete purchases without ever visiting a website.

Conversational AI is actively increasing revenue, with 79% of brands reporting that AI-driven interactions have increased sales and conversion in their business.

The practical implication is that every channel is becoming a storefront. Creating personalized touchpoints with customers earlier in the journey, through proactive engagement, is impacting the bottom line.
Read the full report to explore how AI conversions have increased QoQ by industry.
Pre-purchase hesitation is one of the biggest conversion killers in ecommerce. A shopper lands on your product page, has a question about sizing or compatibility, can't find the answer quickly, and leaves. That's a lost sale that had nothing to do with your product.
Conversational AI changes that dynamic. When a shopper can ask a question and get an accurate, personalized answer in real time, the friction disappears.
Brands using Gorgias saw this play out at scale in 2025. When AI Agent recommended a product, 80% of the resulting purchases happened the same day, and 13% happened the next day.

Brands are further accelerating the buying cycle through proactive engagement. On-site features such as suggested product questions, recommendations triggered by search results, and “Ask Anything” input bars drove 50% of conversation-driven purchases during BFCM 2025.
Explore how AI is collapsing the purchase cycle in Trend 3 of the report.
There's a persistent narrative that AI is making CX teams redundant. The data tells a different story. 62% of ecommerce brands are planning to grow their teams, not cut them. But the scope of those teams is changing.

New roles are emerging around AI configuration and quality assurance. Teams are investing in technical members to write AI Guidance instructions, develop tone-of-voice instructions, and continuously QA results.
CX teams are also bridging the gap between support goals and revenue goals, as the two functions increasingly overlap.

The result is CX teams that are more technical than they were before. Agents who once spent their days answering repetitive tickets are now spending that time on higher-value work: complex escalations, VIP customer relationships, and improving the AI systems and knowledge bases that handle the volume.
Learn more about the evolution of CX roles in Trend #4.
Despite increasing AI adoption, data shows that ecommerce brands shouldn’t strive for 100% automation. Winning brands are building systems in which AI handles repetitive tier-1 tickets, and humans handle complex, sensitive cases.

AI handles speed and scale. It resolves order-tracking requests at 2 a.m., processes return-eligibility checks in seconds, and answers the same shipping question for the thousandth time without compromising quality.
Human agents handle conversations that require context, empathy, or decisions that fall outside the standard playbook. There are several topics where shoppers still prefer human support.

Successful hybrid systems require continuous iteration, meaning reviewing handover topics, Guidance, and reviewing AI tickets on a weekly basis.
Discover how leading brands are balancing human and AI systems in Trend #5.
The 2026 trends are about expansion and standardization. The 2030 predictions are about what comes next.

Voice-based purchasing is the biggest bet on the horizon. Only 7% of brands currently use voice assistants for commerce, but 89% expect it to be standard by 2030. The vision is a customer who can reorder a product, check their subscription status, or manage a return entirely over the phone.
Proactive AI is the other major shift. Rather than waiting for a customer to reach out, AI will anticipate needs based on browsing behavior, purchase history, and where someone is in their relationship with your brand. Think of it as the digital equivalent of a sales associate who remembers what you bought last time and knows what you're likely to need next.
Explore where ecommerce brands are allocating their AI budgets in the full report.
The brands winning in 2026 are creating smart, scalable systems where AIhandles volume and humans handle nuance. They’re treating every conversational channel as an opportunity to serve and sell.
The data is clear: AI adoption is accelerating, customer expectations are rising, and the revenue impact of getting this right is measurable.
{{lead-magnet-1}}

Today, we’re thrilled to announce we’re joining the Shopify Plus Technology partner program.
Over the past few months, we’ve worked with some incredible Shopify Plus merchants like Darn Good Yarn, Fjallraven and Frichti, who serve tens of thousands of customers every month. What they all had in common was a shared commitment to maximizing the efficiency of the customer service team to keep delivering high-quality support as they grow.
We’ve worked with other technology partners, like LoyaltyLion, in order to provide merchants with a holistic view of their customers when they respond to them in an effort to continue delivering best-in-class support interactions. We’ve also worked with the Plus team to leverage the latest features of the Shopify Plus API, to allow agents to create customized solutions for their customers, For example, creating personalized gift cards based on support conversations. Also, check out the guide we wrote comparing Shopify and Shopify Plus for an idea of the additional functionality and benefits ecommerce business owners get when they upgrade to Shopify Plus.
Using our technology, we’re proud to announce that our Shopify Plus customers have managed to improve their support request treatment time by 30%.
By joining the Technology Partner Program, we’re excited to take our collaboration with Shopify Plus and Shopify Plus merchants to the next level, by further enabling more customers to improve their customer service.
"We are glad to welcome Gorgias to the Shopify Plus Technology Partner Program. We’re particularly excited about how they’re helping our merchants provide efficient & personalized customer support, and hope they can help more of them."
Jamie Sutton, Head of Technology Partnerships, Shopify Plus

Great news! Today, we're announcing a new integration with LoyaltyLion. LoyaltyLion is a digital loyalty framework that gives ecommerce stores innovative ways to engage and retain customers.
Our mutual customer Darn Good Yarn uses it to successfully increase customer retention. When they switched from Freshdesk to Gorgias to manage customer support, they wanted to leverage their loyalty program for customer support.
We used their feedback to build the integration with LoyaltyLion, which they have been using for a couple months in beta. Today, we're excited to make it available to all our users.
Here are some of the benefits of this integration:

Overall, this allows your team to use your loyalty program for customer service.
"We love being able to issue our customers loyalty points directly from Gorgias! It's a great way to boost efficiency and also customer retention."
Chloe Kesler, Customer Support manager at Darn Good Yarn
The integration is immediately available on your Gorgias account. If you don't have an account, you can create one here. Then, follow the instructions in our documentation and you can get started!
.avif)
Most customers are loyal to brands because they know what level of service they can expect. As a result, providing an above-average customer experience is key to increase repeat in sales.
It’s relatively easy to provide great support when you get started with your store: your team gets a few dozens of support requests a day, and they respond to them almost instantly. The thing is, this level of service is very hard to maintain as you scale. Response time usually drops, and most brands start using standardized macros to keep up with the pace, which is a poor customer experience.
Your sales have increased, good. Now, you need to get your customer support up to speed

Source: Gorgias customers during the Thanksgiving peak
At Gorgias, we’ve been chatting with 400 stores over the past year, and we’ve seen a lot of them working on crossing this “chasm”.
This post shares learning on how you can build a customer support organization that will scale with your business, and provide best-in-class customer service, which will drive customer retention.
A good place to start is to list the most common reasons customers are contacting you about. Go ahead and manually classify 200 tickets from your support inbox. This should take you about an hour. You can build categories from scratch, or use this spreadsheet of the most common requests for e-commerce companies we built.

Now, you should be able to understand what problems are causing the most pain to customers.
If a specific type of request is above 10% of requests, then it’s a good candidate for optimization. For instance, if you’re getting a lot of “where is my order” questions, here are a few things you can do to deflect those:
Now that you have a good understanding of the reasons customers are contacting you for, you can map the customer journey, and identify what actions your agents need to take to respond to tickets.
Later, you can use this for training purposes, and to identify optimization opportunities.
At Piper, we basically studied the whole customer journey and tried to identify all reasons why someone could contact us (based on previous history). This helped us quickly identify where customers were "blocked"
Finally, let’s analyze the efficiency of your team. Of course, every business is different, but you can use this table to figure out how efficient your agents are compared to other stores. A good metric to track it is ticket closed per month. Just make sure that satisfaction remains consistent.

Related: Learn more about the impact of live chat on sales. And see how Gorgias live chat can help you turn more browsers into buyers with chat campaigns.
Now, let’s work on creating “wow moments” for your customers. If you manage to exceed customers expectations when they contact you, you’re most likely to increase their loyalty and have them refer your store to their friends.
Here are a few ways you can create “wow moments”.
You should be where your customers are. For example, if you have a Facebook page with a large audience, consider it as a real customer support channel. The point is, you should provide the same level of assistance across all support channels that your customer will use.
Also, don't be afraid to contact customers first, especially when they have items in their shopping cart. Offering help or a discount code at the right time could make the difference between a sale and an abandoned cart.
Example: providing high quality support on Facebook
70% of customers consider Facebook as a live chat support. To maximize customer satisfaction, your response time should be no more than 1 min. You’ll then be listed as a very responsive page, which will encourage your customers to respond.

You can also leverage public posts to build relationship with your customers. Another easy way to facilitate customer communication is to remove the need for customers to repeat themselves. On your support platform, make sure your merge Facebook conversations with email tickets. This way, if the customer switches channel, your team will have access to the context of what the customer said before.
Related: Check out our trends and best practices for customer support.
You should leverage every data point you have about the customer to personalize the way you communicate with them:
For how long they have been a customer
Their order preference
Their location
The days of the “we value your business” are over.
Always go an extra mile for your customers. If the customer asks for the status of their order, don’t respond only with the tracking number. Go get the order status on UPS so the customer doesn’t have to do it themselves when they’ll receive your email in the subway with poor network connection.

Another good thing to do is to use a specific tone with your customer, that matches the brand image you want to convey.
If you’re into gifs, you can use them to build a brand tone your own set of gifs, designed for your own brand, and use them in your support emails. You can hire an illustrator on Upwork for that, or build them yourself.
Related: Tips to respond to angry customer emails.
Now that you know the level of support you aim at giving your customers, and you know what actions your agents need to take to get the delivery info, create an RMAs, etc., you can start optimizing the process for them.
To personalize messages, your agents need to have access to customer data. You can leverage the standard Shopify integration from your help desk as a starting point.
Though, it can be relevant to connect other data points to your help desk:
If you’re on Zendesk, enabling the Shopify integration is a good start: it shows how much the customer has spent, and the past orders.
Some Gorgias customers have pretty advanced widgets that display data from Shopify, Stitch & Shipstation. This way, all the customer information is available.
You can create custom widgets for your help desk, so that your agents can trigger actions from your help desk. Here are the most helpful actions:
This is a bit more tricky to implement. You need to build a custom app with buttons that will trigger actions - there are some good tutorials for Zendesk, Freshdesk & Help Scout. At Gorgias, we’ve built integrations with Stitch, Shipstation to embed these actions in the product, and enable you to add your own.
Other 3rd party apps like Chargedesk enable you to refund customers in one click.

Our goal here is to improve the customer experience to drive sales. A good way to track the efficiency of your support work is to compare the behavior of customers that have been in contact with customer support from those who have not.
Shopify helps you easily to this. You can create an integration between your help desk and Shopify to tag customers who reach out to support, using the Shopify API. Say you add a “customer_support” tag to them.

Then, you can use Shopify statistics to monitor how the cohort of customers who have been in touch with your support team behaves, and assess the impact of your efforts with customer support.

Another way to proceed is to tag orders that generated a support tickets. This way, if you work on improving delivery notifications, you can monitor the impact.
Building a scalable support team that provides an amazing customer experience takes time.
Try to test different “wow moments”, iterate on the way you personalize messages, on the tone you’re using, and always track your progress. Among the teams we surveyed, several mentioned they managed to increase sales repeat by 30% after implementing these tactics.
Want to learn more about how customer support can improve your conversion rate and lead to more purchases? Check out our guides to ecommerce upselling and Shopify abandoned cart recovery.

Celery just released their API on Github, currently in beta. Here are some of the cool stuff you can do with it in Gorgias.
When you receive an email from a customer, you can connect your Celery account and see customer information (orders, shipping address, etc.). Here’s what it looks like:

To configure it, grab your Celery access_token, head to integrations, and add an HTTP integration using this URL:
https://api.trycelery.com/v2/orders?buyer.email={ticket.requester.email}
Then you can customize the sidebar to only show the Celery data you need to respond to customers. Click the cog and simply drag and drop elements you want to show.

Celery’s API enables you to perform a few actions from your favorite helpdesk:
Here’s an example of how you can cancel an order from Gorgias itself. Say you already have a macro to cancel an order. Add an HTTP action to it, in this case:
https://api.trycelery.com/v2/orders/{ticket.requester.customer.data[0].number}/order_cancel
Then, when you use this macro and send it to the customer, it will automatically cancel the last order at the same time:

We hope this integration with Celery can save you time. If you'd like to try Celery with Gorgias, shoot us a note! At support@gorgias.com.

TLDR: https://github.com/xarg/pghoard-k8s
This is a small tutorial on how to do incremental backups using pghoard for your PostgreSQL (I assume you’re running everything in Kubernetes). This is intended to help people to get started faster and not waste time finding the right dependencies, etc..
pghoard is a PostgreSQL backup daemon that incrementally backups your files on a object storage (S3, Google Cloud Storage, etc..).
For this tutorial what we’re trying to achieve is to upload our PostgreSQL to S3.
First, let’s create our docker image (we’re using the alpine:3.4 image cause it’s small):
FROM alpine:3.4
ENV REPLICA_USER "replica"
ENV REPLICA_PASSWORD "replica"
RUN apk add --no-cache \
bash \
build-base \
python3 \
python3-dev \
ca-certificates \
postgresql \
postgresql-dev \
libffi-dev \
snappy-dev
RUN python3 -m ensurepip && \
rm -r /usr/lib/python*/ensurepip && \
pip3 install --upgrade pip setuptools && \
rm -r /root/.cache && \
pip3 install boto pghoard
COPY pghoard.json /pghoard.json.template
COPY pghoard.sh /
CMD /pghoard.sh
REPLICA_USER and REPLICA_PASSWORD env vars will be replaced later in your Kubernetes conf by whatever your config is in production, I use those values to test locally using docker-compose.
The config pghoard.json which tells where to get your data from and where to upload it and how:
{
"backup_location": "/data",
"backup_sites": {
"default": {
"active_backup_mode": "pg_receivexlog",
"basebackup_count": 2,
"basebackup_interval_hours": 24,
"nodes": [
{
"host": "YOUR-PG-HOST",
"port": 5432,
"user": "replica",
"password": "replica",
"application_name": "pghoard"
}
],
"object_storage": {
"aws_access_key_id": "REPLACE",
"aws_secret_access_key": "REPLACE",
"bucket_name": "REPLACE",
"region": "us-east-1",
"storage_type": "s3"
},
"pg_bin_directory": "/usr/bin"
}
},
"http_address": "127.0.0.1",
"http_port": 16000,
"log_level": "INFO",
"syslog": false,
"syslog_address": "/dev/log",
"syslog_facility": "local2"
}
Obviously replace the values above with your own. And read pghoard docs for more config explanation.
Note: Make sure you have enough space in your /data; use a Google Persistent Volume if you DB is very big.
Launch script which does 2 things:
#!/usr/bin/env bash
set -e
if [ -n "$TESTING" ]; then
echo "Not running backup when testing"
exit 0
fi
cat /pghoard.json.template | sed "s/\"password\": \"replica\"/\"password\": \"${REPLICA_PASSWORD}\"/" | sed "s/\"user\": \"replica\"/\"password\": \"${REPLICA_USER}\"/" > /pghoard.json
pghoard --config /pghoard.json
Once you build and upload your image to gcr.io you’ll need a replication controller to start your pghoard daemon pod:
apiVersion: v1
kind: ReplicationController
metadata:
name: pghoard
spec:
replicas: 1
selector:
app: pghoard
template:
metadata:
labels:
app: pghoard
spec:
containers:
- name: pghoard
env:
- name: REPLICA_USER
value: "replicant"
- name: REPLICA_PASSWORD
value: "The tortoise lays on its back, its belly baking in the hot sun, beating its legs trying to turn itself over. But it can't. Not with out your help. But you're not helping."
image: gcr.io/your-project/pghoard:latest
The reason I use a replication controller is because I want the pod to restart if it fails, if a simple pod is used it will stay dead and you’ll not have backups.
Future to do:
Hope it helps, stay safe and sleep well at night.
Again, repo with the above: https://github.com/xarg/pghoard-k8s

At Gorgias we recently switched our flask & celery apps from Google Cloud VMs provisioned with Fabric to using docker with kubernetes (k8s). This is a post about our experience doing this.
Note: I'm assuming that you're somewhat familiar with Docker.
The killer feature of Docker for us is that it allows us to make layered binary images of our app. What this means is that you can start with a minimal base image, then make a python image on top of that, then an app image on top of the python one, etc..
Here's the hierarchy of our docker images:
Piece of advice: If you used to run your app using supervisord before I would advise to avoid the temptation to do the same with docker, just let your container crash and let k8s handle it.
Now we can run the above images using: docker-compose, docker-swarm, k8s, Mesos, etc...
There is an excellent post about the differences between container deployments which also settles for k8s.
I'll also just assume that you already did your homework and you plan to use k8s. But just to put more data out there:
Main reason: We are using Google Cloud already and it provides a ready to use Kubernetes cluster on their cloud.
This is huge as we don't have to manage the k8s cluster and can focus on deploying our apps to production instead.
Let's begin by making a list of what we need to run our app in production:
We ran the above in a normal VM environment, why would we need k8s? To understand this, let's dig a bit into what k8s offers:
There are more concepts like volumes, claims, secrets, but let's not worry about them for now.
We're using Postgres as our main storage and we are not running it using Kubernetes.
Now we are running postgres in k8s (1 hot standby + pghoard), you can ignore the rest of this paragaph.
The reason here is that we wanted to run Postgres using provisioned SSD + high memory instances. We could have created a cluster just for postgres with these types of machines, but it seemed like an overkill.
The philosophy of k8s is that you should design your cluster with the thought that pods/nodes of your cluster are just gonna die randomly. I haven't figured our how to setup Postgres with this constraint in mind. So we're just running it replicated with a hot-standby and doing backups with wall-e for now. If you want to try it with k8s there is a guide here. And make sure you tell us about it.
RabbitMQ (used as message broker for Celery) is running on k8s as it's easier (than Postgres) to make a cluster. Not gonna dive into the details. It's using a replication controller to run 3 pods containing rabbitmq instances. This guide helped: https://www.rabbitmq.com/clustering.html
As I mentioned before, we're using a replication controller to run 3 pods, each containing uWSGI & NGINX containers duo: gorgias/web & gorgias/nginx. Here's our replication controller web-rc.yaml config:
apiVersion: v1
kind: ReplicationController
metadata:
name: web
spec:
replicas: 3 # how many copies of the template below we need to run
selector:
app: web
template:
metadata:
labels:
app: web
spec:
containers:
- name: web
image: gcr.io/your-project/web:latest # the image that you pushed to Google Container Registry using gcloud docker push
ports: # these are the exposed ports of your Pods that are later used by the k8s Service
- containerPort: 3033
name: "uwsgi"
- containerPort: 9099
name: "stats"
- name: nginx
image: gcr.io/your-project/nginx:latest
ports:
- containerPort: 8000
name: "http"
- containerPort: 4430
name: "https"
volumeMounts: # this holds our SSL keys to be used with nginx. I haven't found a way to use the http load balancer of google with k8s.
- name: "secrets"
mountPath: "/path/to/secrets"
readOnly: true
volumes:
- name: "secrets"
secret:
secretName: "ssl-secret"
And now the web-service.yaml:apiVersion: v1
kind: Service
metadata:
name: web
spec:
ports:
- port: 80
targetPort: 8000
name: "http"
protocol: TCP
- port: 443
targetPort: 4430
name: "https"
protocol: TCP
selector:
app: web
type: LoadBalancer
That type: LoadBalancer at the end is super important because it tells k8s to request a public IP and route the network to the Pods with the selector=app:web.
If you're doing a rolling-update or just restarting your pods, you don't have to change the service. It will look for pods matching those labels.
Also a replication controller that runs 4 pods containing a single container: gorgias/worker, but doesn't need a service as it only consumes stuff. Here's our worker-rc.yaml:
apiVersion: v1
kind: ReplicationController
metadata:
name: worker
spec:
replicas: 2
selector:
app: worker
template:
metadata:
labels:
app: worker
spec:
containers:
- name: worker
image: gcr.io/your-project/worker:latest
With Kubernetes, docker finally started to make sense to me. It's great because it provides great tools out of the box for doing web app deployment. Replication controllers, Services (with LoadBalancer included), Persistent Volumes, internal DNS. It should have all you need to make a resilient web app fast.
At Gorgias we're building a next generation helpdesk that allows responding 2x faster to common customer requests and having a fast and reliable infrastructure is crucial to achieve our goals.
If you're interested in working with this kind of stuff (especially to improve it): we're hiring!

We've released a new version of the Chrome Extension, with sharing features and a new navigation bar. We hope you'll love it!
Before, the only way to share templates with your teammates was to login on Gorgias.io.
If you're on the startup plan, when you create a template, you can choose who has access to it: either only you, specific people, or your entire team.

The account management section is now available in the extension, under settings.
Tags are now available on the left. It's easier to manage hundreds of templates with them.
You can also navigate through your private & shared templates. Shared templates include templates shared with specific people or with everyone.

We hope you'll enjoy this new version of our Chrome Extension. As usual, your feedback & questions are welcome!

Today, we’re thrilled to announce that we’ve raised a $1.5 million Seed round led by Charles River Ventures and Amplify Partners, to help build our new helpdesk.
We’re incredibly grateful to early users, customers, mentors we’ve met both at and Techstars.
We started the journey with Alex at the beginning of 2015 with our Chrome extension, which helps write email faster using templates. We’ve been pleased all along with customers telling us about how helpful it was, especially for customer support.
While building the extension, we’ve realized that a big inefficiency in support lies in the lack of integration between the helpdesk, the payment system, CRM and other tools support is using. As a result, agents need to do a lot of repetitive work to respond to customer requests, especially when the company is big.

That’s why we’ve decided to build a new kind of helpdesk to enable customer support agents to respond 2x faster to customers. You can find out more and sign up for our private beta here.
When a company has a lot of customers, support becomes repetitive. We want to provide support teams with tools to automate the way they treat simple repetitive requests. This way, they have more time for complex customer issues.
We'll now focus on this helpdesk and on growing the team, oh, and if you'd like to join, we're hiring! We're super excited about this new helpdesk product. If you’re using the extension, don’t worry.
Romain & Alex


