Alfredo Covaleda,
Bogota, Colombia
Stephen Guerin,
Santa Fe, New Mexico, USA
James A. Trostle,
Trinity College, Hartford, Connecticut, USA
This brief piece illustrates how newspapers, TV and radio stations “could” be supplying data/information to their various communities and, at the same time, create data bases that their reporters can draw upon. Why, for example, don't all newspapers have a searchable DB of all possible election results or local ordinances?
President-elect Obama is reportedly going to order the closing of Gitmo [1] as one of his very first acts in office. But it’s far from clear how quickly the coming administration can actually shutter the prison—some guess it will take the better part of a year. A key factor, of course, will lie in just who the 248 detainees being still held [2] at Gitmo are and what exact the government believes they’ve done.
One of the handiest—and least-noticed—places for finding that information is the New York Times’ “Guantanamo Bay Docket [2].”
Launched quietly on November 3 (when some folks were busy focusing on other impending news [3]), the searchable database has government documents on all of the prisoners the Pentagon has acknowledged have been held at Gitmo. 779 have been detained there since 2002; 248 remain there today.
For example, here are the details of the government’s case [4] against alleged 9/11 mastermind Khalid Shaikh Mohammed. (KSM is just one of 16 high-value detainees still held at Gitmo. The Times’ project has details on each one [5].)
The docket has about 16,000 pages overall, almost all from court hearings. You can search through all of the documents. For example, interested in seeing all the times “torture” has been mentioned [6] at a hearing? There’s also a handy timeline [7].
We found the project so interesting that we decided to ring Margot Williams, the Times’ database research editor, who has spearheaded the effort. Margot has been involved in breaking Gitmo stories for years. In fact, she’s such a junkie, she said she put a recording of KSM’s confession [8] on her cellphone.
“I don’t know what compels me to do this, but for years I’ve been compiling lists of people in secret detention. So I’ve been in ‘list mode’ for seven years,” Williams told us.
What the Times did was combine the information Williams has been compiling and linked it up with the government documents on each detainee. “Wikipedia and a few other sites have posted some documents. But they’re not searchable.”
Williams warns that the documents—even the thousands of pages of them—don’t tell anything close to the full story. “This is only one side, it’s not presenting the defense teams’ sides. I’m hoping to get those documents in, but that’s hard because most of them haven’t been posted.
“I’ve read the 16,000 pages. That’s where I pull the information that goes in the database. I am a reader of documents. There’s no substitute for reading every line of every page. As they said on the 9/11 commission report, it’s about putting together the mosaic and seeing the patterns.
“In my mind’s eye, this is the form I’d hope the government would make them available in rather than dumping them in unusable formats. I’m glad we’ve made the documents more useful for people to figure out for themselves who these [prisoners] are.”
http://www.propublica.org/article/gitmo-database-details-779-prisoners-cases
Nathan, over at Flowingdata.com, posts this interesting data visualization from the Baylor College of Medicine. No, it probably doesn't give a science writer a story in itself, but the concept of taking a complex data set and illustrating that data with the right tool — in this case, Circos — good generate some interesting reporting vectors. For example, could Circos show us something about traffic patterns? Ambulance or fire department response times? We're not sure, but we hope someone could probe this a bit.
The thing about cancer cells is that they suck. Their DNA is all screwy. They've got chunks of DNA ripped out and reinserted into different places, which is just plain bad news for the cells in our body that play nice. You know, kind of like life. Researchers at the Baylor College of Medicine in Houston have compared the DNA of a certain type of breast cancer cell to a normal cell and mapped the differences (and similarities) with the above visualization.
The graphic summarizes their results. Round the outer ring are shown the 23 chromosomes of the human genome. The lines in blue, in the third ring, show internal rearrangements, in which a stretch of DNA has been moved from one site to another within the same chromosome. The red lines, in the bull's eye, designate switches of DNA from one chromosome to another.
Some design would benefit the graphic so that your eyes don't bounce around when you look at the technicolor genome but it's interesting nevertheless.
Check out the Flare Visualization Toolkit or Circos if you're interested in implementing a similar visualization with the above network technique.
http://flowingdata.com/2008/12/29/researchers-map-chaos-inside-cancer-cell
From O'Reilly's Radar:
GeoData Explorations: Google's Ever-Expanding Geo Investment
Posted: 16 Dec 2008 02:40 PM CST
Google has been investing lots of money in geodata acquisition. Some of the money is being spent externally: they've inked an exclusive satellite imagery deal with GeoEye (Radar post) and a data sharing deal Tele Atlas (Radar post). And some is being spent internally with Mapmaker, Street View and the web. Over the past week Google has been sharing visualizations of their internally gathered geodata. Here's a round-up of them.
The image above was released on December 9th. It shows how much of the US is available via Street View. According to the post Street View imagery increased 22 fold around the world in 2008.
The dark image above was released on December 11th. It highlights the parts of the world that are being mapped on Google's Mapmaker by users (Radar post). Mapmaker is now live in 164 countries. According to the map it has gained the most traction in Africa and the Indian sub-continent. The Google Mapmaker team has released timelapse videos of Mapmaker building cities on the Mapmaker YouTube Channel. I've embedded one after the jump.
This final image shows all the points described by GeoRSS and KML all over the world. It was shown at Where 2.0 2007 by Michael Jones (video). Unsurprisingly, this image and the Mapmaker image show opposite data density concentrations.
In some more GeoData Explorations posts this week I will look at OSM vs Google and some surprising trends in KML.
This timelapse video of Da Lat, Vietnam being created is one of the most impressive. It shows a town and lake emerge from a blank slate.
Colleague Owen Densmore points us to this page with these comments:
This use of gps may play a role in understanding economic development in any city by watching the flows within the city: http://digitalurban.blogspot.com/2008/12/gps-city-tracks-1-year-in-24-hours-via.html This gets me to an aspect of ED I'm interested: MicroED. It comes from the observation that all cities' ED is unique. Think about every city you've lived in and you'll notice that each was unique. For me, Rochester NY: Kodak/Xerox company towns; Silicon Valley: A network of startups and established companies with a highly mobile social/skill network. Here in Santa Fe, we are similarly unique. I think this is core: discover your unique environment and capitalize on improving it through managing it. Data farming your city. Graph its flows. — Owen
This use of gps may play a role in understanding economic development in any city by watching the flows within the city: http://digitalurban.blogspot.com/2008/12/gps-city-tracks-1-year-in-24-hours-via.html
This gets me to an aspect of ED I'm interested: MicroED. It comes from the observation that all cities' ED is unique. Think about every city you've lived in and you'll notice that each was unique. For me, Rochester NY: Kodak/Xerox company towns; Silicon Valley: A network of startups and established companies with a highly mobile social/skill network. Here in Santa Fe, we are similarly unique.
I think this is core: discover your unique environment and capitalize on improving it through managing it. Data farming your city. Graph its flows.
— Owen
GPS tracks can show the 'life' of a city, which parts of the city are working, areas that are no go zones and sections dedicated to shopping, work, entertainment etc. The possibilities for using GPS data to examine our cities 'health' are intriguing which turns us to the work by Fabian over at http://urbantick.blogspot.com/
The movie below illustrates Fabian's paths around the city of Plymouth over 365 days, compressed and visualised in Google Earth:
Google Earth is an excellent tool for displaying GPS data, especially over time, we are just starting to look into other options, perhaps After Effects – any thoughts or ideas for visualising GPS tracks over time would be great…
See http://urbantick.blogspot.com/ for more movies and examples on visualising GPS tracks in the city.
Another good pointer from Nathan at Flowing Data:
UUorld (pronounced “world”) is a 4-dimensional mapping tool that lets you explore geographic data – the fourth dimension being time. The interface will remind you a bit of Google Earth with the map, pan, zoom, etc, however, UUorld isn't trying to replace Google Earth. In fact, it'll probably be better if you use it with Google Earth. Think of it as another tool to add to your box of mapping toys.
UUorld's focus is on finding trends over space and time. Load your own data or import data from UUorld's data portal, and then play it out over time. Spatial boundaries undulate up and down as land masses look a bit like skyscrapers. Color and boundary lines are customizable. When you're satisfied with the results, record it as video or export as KML, and then import into Google Earth or whatever else you want.
How effective is this method of visualization though? There's the usual argument of area perception, but does color-coding and vertical dimension make up for that? Discuss amongst yourselves.
To see the relevant data of the “International Survey of Journalism Educators in the Digital Age,” click here.
Yes, yes. We know it's a promo piece for ESRI and the utility of GIS. But the article below from ESRI's ArcNews also highlights the company-wide utility of GIS for journalists, at least those who understand that what we do requires a wider vision than just writing well. Additionally, the article illustrates in-house expertise that folks in the newsroom could consider tapping
The Star Tribune newspaper of Minneapolis-St. Paul, Minnesota, is using ArcLogistics GIS software to route its delivery drops. It is projected to save the newspaper more than a half million dollars in the next five years. By employing ArcLogistics in its efforts to serve the Minneapolis-St. Paul metropolitan region, greater Minnesota, and surrounding states, the Star Tribune has found that using GIS technology for intelligent routing has delivered an economic advantage to the company.
Given current economic trends, the company began looking for new areas where the newspaper could save money. Traditionally, the newspaper used wall maps with pushpins to determine delivery routes for single-copy papers. Rerouting was a time- and labor-intensive activity that required three employees from multiple departments to sit in the map room for four hours a day over the course of a month. Determining new routes is a necessity each time one of the seven advertisement zone boundaries is shifted or when a threshold amount of new or removed newspaper drop locations is reached. After reviewing several options to increase the efficiency of delivery routes, the Star Tribune partnered with Truck Dispatching Innovations, an ESRI Business Partner from Chicago, Illinois, to implement ArcLogistics. After a two-week startup period geocoding more than 3,700 delivery drop points and the routes of 39 trucks, employees used GIS to create new routes. These outcomes had many benefits. Using this new methodology, one staff member inputs a list of delivery route changes into ArcLogistics and, in half the time of the traditional method, creates and shares maps displaying new routes. The staff perform what-if scenarios, such as including different ad-zoned papers on the same truck. These reveal route options that could further increase the efficiency of delivery routes.
The Star Tribune expects a payback on its investment in 2.5 months and a five-year net savings of $672,740. This positive return on investment provides evidence showing the success of the venture to the company's financial director. The Star Tribune analyzed the benefits of its investment in ArcLogistics by measuring fixed and variable costs. Fixed costs include the lease price and maintenance expenses for delivery trucks, as well as the initial cost and yearly maintenance of the software. One variable cost is the number of miles driven, which determines gas costs per route. Another variable cost is the number of hours driven, which determines the wage cost per driver per route. Fixed costs are added to variable costs to determine route costs. Combining cost savings in these four areas shows a more accurate cost savings, rather than just looking at the savings as stand-alone figures. The Star Tribune found significant savings of route costs, including the number of trucks needed, miles driven, and time spent delivering newspapers.
For more information, contact Al Olson, fleet logistics manager, Star Tribune (e-mail: aolson@startribune.com), or John Handler, principal, Truck Dispatching Innovations (e-mail: jhandler@TDInnovations.com). To learn more about ArcLogistics, visit www.esri.com/arclogistics.
Amy Disch, library director of The Columbus (Ohio) Dispatch, sends along these links via the News Librarians' listserv (newslib@listserv.unc.edu). This is a gentle reminder about how the foundations of good publications today rest, first, on the integration of library AND IT skills.
Watch them in the order listed:
IAJ co-founder Steve Ross has long argued that Craig's List HAS NOT contributed in a major way to the decline of North American newspaper advertising revenues. Here's his latest analysis:
This isn't rocket science. Craig's List had NOTHING to do with a supposed decline in classified ad revenue. Here's the raw PRINT classified revenue data, right off the NAA website. (If anyone doesn't use excel 2007 I can send the data file in another format, but everyone should be able to read the chart as a jpg). Click here for bar chart Note that the big change that pushed classified ad volume up in the 90s was employment advertising. Damn right. The country added 30 million new jobs in that period, and the number of new people entering the workforce declined because births had declined in the mid-1970s. More competition for bodies = more advertising needed. Knock out the employment data and everything else stayed steady or INCREASED for newspaper classified. The past 7 years were not as good for employment ads, but still better than in pre-web days. There was indeed sharp deterioration in 2007 (and of course, 2001), as the economy soured. There are some missing data (idiots) right around the time the web came in — 1993-4. But just look at 1994-2006 — the “web years.” Total print classified ad dollar volume was $12.5 billion in 1994, $17 billion in 2006, roughly in line with inflation AT A TIME WHEN CIRCULATION FELL and even newspapers managed to get some new online revenue!!! Look, I can do this with ad lineage (which didn't rise much at all but stayed ahead of circ declines), I can compare with display ad figures, I can do Craig's List cities vs non-Craig, I can add back the web revenue because in fact newspapers allocate revenue wrong, to preserve existing ad sales commission schemes, and thus undercount web revenue. I can do ad revenue per subscriber. And on and on. All those corrections make this look even better for newspapers. This is SO OBVIOUS that I just do not understand the “Craigs List has killed us” argument or even the”web killed us” argument. It is (to me, anyway) a transparent lie. Either the newspaper barons are so inanely stupid that they don't understand their own business, or they are incompetent managers, looking for an excuse. Maybe both. But oddly enough, Craig Newmark believes he did the damage. I've been on several panels with him where he has apologized for killing newspapers. I might also add that some obviously web-literate societies are seeing a newspaper boom. Germany is an example.
This isn't rocket science. Craig's List had NOTHING to do with a supposed decline in classified ad revenue.
Here's the raw PRINT classified revenue data, right off the NAA website. (If anyone doesn't use excel 2007 I can send the data file in another format, but everyone should be able to read the chart as a jpg). Click here for bar chart
Note that the big change that pushed classified ad volume up in the 90s was employment advertising. Damn right. The country added 30 million new jobs in that period, and the number of new people entering the workforce declined because births had declined in the mid-1970s. More competition for bodies = more advertising needed.
Knock out the employment data and everything else stayed steady or INCREASED for newspaper classified.
The past 7 years were not as good for employment ads, but still better than in pre-web days.
There was indeed sharp deterioration in 2007 (and of course, 2001), as the economy soured.
There are some missing data (idiots) right around the time the web came in — 1993-4.
But just look at 1994-2006 — the “web years.” Total print classified ad dollar volume was $12.5 billion in 1994, $17 billion in 2006, roughly in line with inflation AT A TIME WHEN CIRCULATION FELL and even newspapers managed to get some new online revenue!!!
Look, I can do this with ad lineage (which didn't rise much at all but stayed ahead of circ declines), I can compare with display ad figures, I can do Craig's List cities vs non-Craig, I can add back the web revenue because in fact newspapers allocate revenue wrong, to preserve existing ad sales commission schemes, and thus undercount web revenue. I can do ad revenue per subscriber. And on and on.
All those corrections make this look even better for newspapers.
This is SO OBVIOUS that I just do not understand the “Craigs List has killed us” argument or even the”web killed us” argument.
It is (to me, anyway) a transparent lie. Either the newspaper barons are so inanely stupid that they don't understand their own business, or they are incompetent managers, looking for an excuse. Maybe both.
But oddly enough, Craig Newmark believes he did the damage. I've been on several panels with him where he has apologized for killing newspapers.
I might also add that some obviously web-literate societies are seeing a newspaper boom. Germany is an example.