There's nothing we enjoy more than a sweet data viz. They're a great way to take a lack luster table of data and dress is up to make it something truly special. To show off what some of our favourite viz partners can do, we've compiled these 5 awesome data vizzes for you to explore.
1. Add some sparkle with Silk
The holidays are here and that means it's time to pop open some bubbly! To help you choose just the right sparkling wine to liven up your dinner party, our friends at Silk.co have created this awesome data viz.
2. Sing along with Tagul
Nothing says Christmas to me like singing carols. Here's a simple word cloud I made with Tagul that shows what the most popular words are in Christmas carols (excluding the word Christmas).
3. Make your list and check it twice with Tableau
Jewel Loree used Tableau least year to find out which celebs would make Santa's nice and naughty lists.
4. Give the gift of love with Infogr.am
Keeping with the celebrity theme, our own Sally Hadidi used Infogr.am to visualize how much each of these popular celeb duos is worth!
5. Get ready for the New Year with Plot.ly
if you're thinking about making a fresh start in the New Year, check out this Plot.ly viz on which careers are most popular in SF.
We all know that seeing is believing. So this Christmas, don't just take our word for it; see real life uses for data that import users have come up with. We're featuring them on a new page called #DataPunk Showcase. These are stories to help inspire you and help get those creative data juices flowing. We've got everything from fun data vizzes, to lead generation, to apps and so much more. We're starting you off with the 6 best ones, but we'll be adding more soon!
Have a cool use case? Tell us about it! You could be featured in our showcase. Just email us at firstname.lastname@example.org.
A guest post by Federico Pascual, Business Development Manager at MonkeyLearn
Ever wondered which city has the most arts jobs? Or recruitment openings? Will your career be better off in NYC or SF? Turns out you can answer these and many more questions by doing some simple data analysis. Using three awesome free tools (import.io, MonkeyLearn and Plot.ly) you can obtain, categorize and visualize all the data we need in just 10 minutes!
In this tutorial, we’ll start by using import.io’s Magic API to obtain hundreds of job postings from LinkedIn. With this, we’ll write a small Python script to process and send the data to MonkeyLearn so they may be categorized using one of the public modules, effectively enriching the original data with the category each job belongs to. We’ll then finish by uploading it all to Plot.ly and obtaining some insights from the data through the use of graphs.
Let’s get started!Read More
A guest post by Joe Warbington (Solutions Architect, Healthcare at Qlik)
Recently, Inc released their latest list of the fastest-growing private U.S. companies. "The list is compiled by measuring each company's percentage revenue growth over a four-year period. As a data set, it can tell you some pretty interesting things such as which industries are growing the fastest and which cities high-growth companies are gravitating towards." according to import.io.
California tech companies dominate the 2014 list, perhaps as expected. Curious to see how Wisconsin-based companies performed, I took the list and quickly added it to Qlik Sense - a next-gen data visualization and discovery application.
In minutes, I was able to to see the 70 Wisconsin companies that made the list this year. Quite a few healthcare-IT related firms help boost the workforce (17,984 people) and overall revenue generated ($7 billion USD):Read More
One of my absolute favorite things to do with the import.io tools is to build mixes. A mix is just a bunch of connectors that all take the same input. To show you a quick example, I've built 7 connectors to sites which sell Christmas jumpers (cause who doesn't love a good Christmas jumper), then combined them all into one data set with a single search box!
For more on how to build mixes, check out this tutorial.
Usage. It’s the ultimate driver of every business owner and CEO. We obsess over it. I have an entire 42” TV hanging over my desk with a real-time dashboard of numbers that measure this usage in different ways. I can access this dashboard on my laptop (where it is bookmarked), my tablet and even an app on my iPhone. It’s often the first thing I look at in the morning and sometimes the last thing I look at before I go to bed - much to my wife’s annoyance. Like I said, it’s an obsession.
And it’s not just because I’m a data nerd. Metrics matter. More accurately, usage matters. The more people use your platform, the more benefit they’ll get out of it, the more likely they are to tell others about it, and the more likely your product is to go viral. Usage is the key to success - and in tech, more is always better.Read More
Up until now, we've been mostly talking about ways to get your data. But that's only the first half of the Datamas story! Once you've got it, you need to do something with it. And if that doing means integrating your data over the API, then you're in luck! We've got 8 client libraries to help you out, full of example queries and step-by-step guides.
Here they are in all their glory:
A crawler is the perfect tool to help you get a huge amount of static data from an entire website. You simply train it on 5 pages as examples and then let it go! The crawler will follow the links on your example pages to find other pages with similar data and extract that!
Because crawlers can sometimes take a long time, we've given you 9 advanced options to help make them as efficient as possible:
- Page depth - how many "clicks" from the start pages the crawler will travel looking for data
- Save log - a file of all the URLs that have been visited and which ones have been converted or failed
- Save stream - creates a file with the data that has been converted thus far
- Simultaneous pages - the number of pages the crawler will visit at one time
- Pause between pages - how long the crawler will wait before going to the next link
- Where to crawl - where the crawler will journey to find new links -
- Where not to crawl - links you don’t want the crawler to visit
- Where to extract data from - the URL pattern generated from your example pages
- Enable cookies - run your crawler with cookies enabled
For more on how to build a crawler and the different options we offer, check out our famous Crawler Webinar.
This latest webinar is another classic, but always popular, Getting Started. But instead of crawling for jeans on Asos (you regulars will know what I’m on about), I thought I’d take inspiration from the impending holiday season and make it a bit more festive!
Nothing says Christmas to me more than jumpers and delicious food, so that’s what I’ll be focusing on...apologies in advance for making you hungry :-).Read More
Here at import.io, we are always working to improve our service and tools. A big part of that improvement, is getting feedback and ideas from you guys. If you haven't already you should check out the ideas forum, where you can submit your ideas for new features/improvements and vote for the ones you want us to develop first. You get 10 votes to spend on whatever you want (you can award each idea up to 3 votes)!
Every week, David (CEO) and Matt (CTO) sit down and go through the ideas forum to see what the most popular features are to help decide what to put in the development road map. Once an idea has been slated for development we'll notify you in the ideas forum and keep you updated on its progress, so you'll be the first to know when it is launched.
To prove that we really do listen, here's a list of the top 10 ideas that have already been built into features:
- Authenticated APIs - allow the tool to access data behind a login
- Command line crawler - ability to schedule crawlers and send crawled data directly to 3rd party location
- Where to crawl - specify a lit of URLs to crawl
- Source code view - see the underlying HTML of the page you're extracting
- Toggle CSS - ability to turn CSS off
- Log file - have access the the log file when crawling
- Crawler snapshot - for data recovery
- Default column value - be able to set a static value in a column
- Source URL - ability extract to extract the URL the data came from
Have an idea? GREAT! We want to know what it is, so head over to the ideas forum and tell us about it.
And don't forget to cast your 10 votes for your favorite ideas!