Monitoring InfluxDB with Scout

By Derek Bullet_white Posted in App Monitoring Bullet_white Comments Comments

We're using InfluxDB in our new app monitoring service.

While InfluxDB hasn't reached 1.0 yet, it has loads of potential and has been holding up well during our BETA period. Don't worry, we'll talk more about InfluxDB in coming posts.

So, how are monitoring InfluxDB performance? Here's how we get an overview of our app performance. InfluxDB is one of the categories we track:


When there's a slow request, we can dig into details, including viewing the actual InfluxDB query:


Need some InfluxDB monitoring action?

Sign up for early access and ping us at to let us know you are using InfluxDB.


Introducing easy StatsD with Scout

By Mark Bullet_white Comments Comments

The easiest way to unleash StatsD.

One agent. Minimal overhead. Robust language support. A unified monitoring solution for your servers and metrics.

We've added StatsD support to our monitoring agent. With Scout, you are just minutes away from StatsD-backed charts and alerts. Use StatsD to report code execution times, user signup rates, and more.

StatsD generated metrics are first class citizens in Scout - coexisting with every metric available via Scout.

Don't just take our word for it, Scout customer Martin Kelly had this to say:

It took about an hour to get all of our prod and preprod custom metrics into Scout and displayed on our dashboard. Both the client and project manager were very happy!"

Why StatsD + Scout?

Prior to today, all metrics in Scout were generated by our agent (monitoring system resource usage) or plugins (monitoring services). This works great for sampling metrics, but it's not a great fit for event-based metrics (ex: tracking user signups, response times, etc).

StatsD is a great fit for event-based metrics, but rolling your production-grade setup for StatsD is involved. We also want to see all of our metrics (and configure alerting) from a single app.

We tested StatsD+Scout internally first, loved it, and rolled out to customers during a preview stage. Today, StatsD is battle-tested and ready for your metrics.

Quick tip: replacing metric logging with StatsD

First time with StatsD? Here's a tip: if you are logging a metric, it probably makes more sense to send it StatsD.


logger.warn "Error Occurred!"



...which gives you ready-to-go charts and alerting (ex: alert when the error rate exceeds 50 errors/min).

Get started with StatsD today!

Additional resources to get you started:

Questions? Comments? Contact us at

More updates on StatsD? Follow us on Twitter.


The making of app monitoring: the health dashboard

By Derek Bullet_white Posted in App Monitoring Bullet_white Comments Comments

We're battle-scarred devs building the focused app monitoring service we've always wanted. We're blogging about the adventure below.

Customers telling me our app is slow? I'm looking at a response time graph.

On the front page of Hacker News? I'm looking at requests per-second and response time on a graph.

Lots of things going wrong? Show me ALL the metrics.

The challenges with building a one-page dashboard of app health?

  • What's important to me today might not be tomorrow
  • I need to see all key metrics at once to ensure I'm not missing a correlation (ex: spike in response time and error rates)
  • Ability to magnify a metric on a chart for more details

The first step is admitting I have a problem

We track eight key health metrics for our applications:

  • Response Time by category (time spent in Ruby, Postgres, Elasticsearch, etc)
  • Throughput
  • Error Rate
  • Apdex
  • Capacity % (the utilization of our application worker processes)
  • App Instances (how many processes are serving our app across all of our nodes)
  • CPU Usage % (average cpu usage of the app on each node)
  • Memory Usage (average memory usage of the app on each node)

So, what are some approaches to help me get an at-a-glance view of app health?

Read More →


Reversing the GoDaddy-ification of application monitoring

By Derek Bullet_white Posted in App Monitoring Bullet_white Comments Comments

Scout is an "oops" company. We didn't build our product with the intention of turning it into a company, but we certainly can't imagine life without it today.

Scout was started out of frustration. The prospect of setting up and using a Nagios-like server monitoring solution was so terrifying, we'd rather build our own. We built a simple monitoring agent and an accompanying Rails-backed web interface and used it monitor our own apps. When debugging performance issues, we started sharing access to the app with our hosting provider, Rails Machine. They loved it and started using it.

We really enjoyed building the product, put a price tag on it, and over a bit of time, it became our full-time thing.

That frustration is back: app monitoring

Application performance monitoring (APM) products have a tendency to evolve into a GoDaddy-like experience.

The tools for monitoring apps are continually becoming more complex and difficult to use. It's the second law of thermodynamics applied to software. There's an ever increasing tendency toward disorder.

It's time to reset application monitoring.

From our own frustrations and those our customers have shared with us, it's clear app monitoring needs a craft brew-alternative: a polished, focused take on application monitoring. A product focused on solving performance issues as fast as possible and not overwhelming you with clutter.

We're building the craft brew of app monitoring

A bit ago, we decided to build an app monitoring product. We've got an awesome team dedicated to it and it's coming along fast. There's some core beliefs we're staring with:

  • Support multiple languages and frameworks. We know from experience that we're mixing together more languages and frameworks than ever before. It's key to view their performance from a single interface.
  • Easy time range diffs. The UI must be built to make it easy to compare deploys, config changes, or general trends as an application ages.
  • Context. How is performance for our highest-paying customers? Is a performance issue impacting everyone or a subset of customers? Is slowness primarily associated with one database node? Make it easy to apply the context that matters to you.
  • Aggregrate what's slow. We learn a lot from investigating slow requests. Rather than paging through metrics on individual slow requests, aggregate the call stacks of slow requests together. Apply the context from above. Know with certainty that an endpoint is slow because of a specific query for X% of your customers.

Sign up for our BETA

Get yourself on our early access list. We'll be inviting folks into APM ahead of our October launch. It's a great time to help shape the direction of Scout APM.

More to come

We'll be blogging about the product dev process right here, starting with the design decisions behind our application health dashboard:

Scout APM

Follow us on Twitter for the highlights and signup for early access.


The Curious Case of the StatsD Timer

By Mark Bullet_white Posted in HowTo Bullet_white Comments Comments


Instrumenting our application with StatsD is easy, especially when we just stick to Counters and Gauges. These metrics return just a single value when implemented. When you get to Timers, however, StatsD steps up it's game and returns eight metrics.

So let's explore the curious case of the timing metric. What do all these metrics mean? How can we use this for instrumenting our application?

Read More →


New Dog in the Scout Kennel

By Mark Bullet_white Comments Comments

Please join us in welcoming Chris Schneider to the Scout team!

We've had our eye on Chris for awhile now, and we couldn't be more excited about him joining the team. Nearly a Fort Collins native (rare around these parts), Chris joins us with over 15 years of development experience - with the majority of his time spent in mission critical Rails apps. He's a programmer's programmer. Chris has hosted the Coding Hangout in Fort Collins for over a year now. Chris loves to teach and we're all blown away by his mad VIM skills.

Chris will be our lead-developer on our new Scout Application Monitoring service,. His insight and experience will be invaluable in making this product a world-class application monitoring solution.

In his free time, Chris can normally be found crushing trivia at a local tavern, banging out some Haskell or playing games with friends.

We all look forward to working with Chris!


Older posts: 1 ... 15 16 17 18 19 ... 68