US Government and Proprietary Databases: Will Procurement Roadblocks Get Set Up before October 1, 2015?

July 20, 2015

I don’t do the government work stuff anymore. Too old. But some outfits depend on the US government for revenue. I should write “Depend a lot.”

I read “Why Government Needs Open Source Databases.” The article is one of those which is easily overlooked. With the excitement changing like a heartbeat, “database” and “government” are not likely to capture the attention of the iPhone and Android crowd.

I found the article interesting. I learned:

Open source solutions offer greater flexibility in pricing models as well. In some cases, vendors offering open source  databases price on a subscription-based model that eliminates the licensing fees common to large proprietary systems. An important element to a subscription is that it qualifies as an operating expense versus a more complex capital expenditure. Thus, deploying open source and open source-based databases become a simpler process and can cost 80 to 90 percent less than traditional solutions. This allows agencies to refocus these resources on innovation and key organizational drivers.

Wow, cheaper. Maybe better? Maybe faster?

The article raises an interesting topic—security. I assumed that the US government was “into” security. Each time I read disinformation about the loss of personnel data or a misplaced laptop with secret information on its storage device, I am a doubter.

But the article informs me:

Data security has always been and will continue to remain a major priority for government agencies,  given the sensitive and business-critical nature of the information they collect. Some IT departments may be skeptical of the security capabilities of open source solutions. Gartner’s 2014 Magic Quadrant for Operational Database Management Systems showed that open source database solutions are being used successfully in mission-critical applications in a large number of organizations. In addition, mature open source solutions today implement the same, if not better, security capabilities of traditional infrastructures. This includes SQL injection prevention, tools for replication and failover, server-side code protections, row-level security and enhanced auditing features, to name a few. Furthermore, as open source technology, in general, becomes more widely accepted across the public sector – intelligence, civilian and defense agencies across the federal government have adopted open source – database solutions are also growing with specific government mandates, regulations and requirements.

I knew it. Security is job one, well, maybe job two after cost controls. No, no, cost controls and government activities do not compute in my experience.

Open source database technology may be the horse the government knights can ride to the senior executive service. If open source data management systems get procurement love, what does that mean for IBM and Oracle database license fees?

Not much. The revenue comes from services, particularly when things go south. The license fees are malleable, often negotiable. The fees for service continue to honk like golden geese.

Net net: Money will remain the same, just be taken from a different category of expense. In short, the write up is a good effort, but offers little in the way of bad news for the big database vendors. On October 1, 2015, not much change in the flowing river of government expenditures which just keep rising like the pond filled with mine drainage near my hovel in Kentucky.

Stephen E Arnold, July 20, 2015

Publishers Out Of Sorts…Again

July 20, 2015

Here we go again, the same old panic song that has been sung around the digital landscape since the advent of portable devices: the publishing industry is losing money. The Guardian reports on how mobile devices are now hurting news outlets: “News Outlets Face Losing Control To Apple, Facebook, And Google.”

The news outlets are losing money as users move to mobile devices to access the news via Apple, Facebook, and Google. The article shares a bunch of statistics supporting this claim, which only backs up facts people already knew.

It does make a sound suggestion of traditional news outlets changing their business model by possibly teaming with the new ways people consume their news.

Here is a good rebuttal, however:

“ ‘Fragmentation of news provision, which weakens the bargaining power of journalism organisations, has coincided with a concentration of power in platforms,’ said Emily Bell, director of the Tow Center at Columbia university, in a lead commentary for the report.”

Seventy percent of mobile device users have a news app on their phone, but only a third of them use it at least once a week. Only diehard loyalists are returning to the traditional outlets and paying a subscription fee for the services. The rest of the time they turn to social media for their news.

This is not anything new. These outlets will adapt, because despite social media’s popularity there is still something to be said for a viable and trusted news outlet, that is, if you can trust the outlet.

Whitney Grace, July 20, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Hadoop Rounds Up Open Source Goodies

July 17, 2015

Summer time is here and what better way to celebrate the warm weather and fun in the sun than with some fantastic open source tools.  Okay, so you probably will not take your computer to the beach, but if you have a vacation planned one of these tools might help you complete your work faster so you can get closer to that umbrella and cocktail.  Datamation has a great listicle focused on “Hadoop And Big Data: 60 Top Open Source Tools.”

Hadoop is one of the most adopted open source tool to provide big data solutions.  The Hadoop market is expected to be worth $1 billion by 2020 and IBM has dedicated 3,500 employees to develop Apache Spark, part of the Hadoop ecosystem.

As open source is a huge part of the Hadoop landscape, Datamation’s list provides invaluable information on tools that could mean the difference between a successful project and failed one.  Also they could save some extra cash on the IT budget.

“This area has a seen a lot of activity recently, with the launch of many new projects. Many of the most noteworthy projects are managed by the Apache Foundation and are closely related to Hadoop.”

Datamation has maintained this list for a while and they update it from time to time as the industry changes.  The list isn’t sorted on a comparison scale, one being the best, rather they tools are grouped into categories and a short description is given to explain what the tool does. The categories include: Hadoop-related tools, big data analysis platforms and tools, databases and data warehouses, business intelligence, data mining, big data search, programming languages, query engines, and in-memory technology.  There is a tool for nearly every sort of problem that could come up in a Hadoop environment, so the listicle is definitely worth a glance.

Whitney Grace, July 17, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

The Skin Search

July 15, 2015

We reported on how billboards in Russia were getting smarter by using facial recognition software to hide ads advertising illegal products when they recognized police walking by.  Now the US government might be working on technology that can identify patterns on tattoos, reports Quartz in, “The US Government Wants Software That Can Detect And Interpret Your Tattoos.”

The Department of Justice, Department of Defense, and the FBI sponsored a competition that the National Institute of Standards and Technology (NIST) recently held on June 8 to research ways to identify ink:

“The six teams that entered the competition—from universities, government entities, and consulting firms—had to develop an algorithm that would be able to detect whether an image had a tattoo in it, compare similarities in multiple tattoos, and compare sketches with photographs of tattoos. Some of the things the National Institute of Standards and Technology (NIST), the competition’s organizers, were looking to interpret in images of tattoos include swastikas, snakes, drags, guns, unicorns, knights, and witches.”

The idea is to use visual technology to track tattoos among crime suspects and relational patterns. Vision technology, however, is still being perfected.  Companies like Google and major universities are researching ways to make headway in the technology.

While the visual technology can be used to track suspected criminals, it can also be used for other purposes.  One implication is responding to accidents as they happen instead of recording them.  Tattoo recognition is the perfect place to start given the inked variety available and correlation to gangs and crime.  The question remains, what will they call the new technology, skin search?

Whitney Grace, July 15, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Does America Want to Forget Some Items in the Google Index?

July 8, 2015

The idea that the Google sucks in data without much editorial control is just now grabbing brain cells in some folks. The Web indexing approach has traditionally allowed the crawlers to index what was available without too much latency. If there were servers which dropped a connection or returned an error, some Web crawlers would try again. Our Point crawler just kept on truckin’. I like the mantra, “Never go back.”

Google developed a more nuanced approach to Web indexing. The link thing, the popularity thing, and the hundred plus “factors” allowed the Google to figure out what to index, how often, and how deeply (no, grasshopper, not every page on a Web site is indexed with every crawl).

The notion of “right to be forgotten” amounts to a third party asking the GOOG to delete an index pointer in an index. This is sort of a hassle and can create some exciting moments for the programmers who have to manage the “forget me” function across distributed indexes and keep the eager beaver crawler from reindexing a content object.

The Google has to provide this type of third party editing for most of the requests from individuals who want one or more documents to be “forgotten”; that is, no longer in the Google index which the public users’ queries “hit” for results.

According to “Google Is Facing a Fight over Americans’ Right to Be Forgotten.” The write up states:

Consumer Watchdog’s privacy project director John Simpson wrote to the FTC yesterday, complaining that though Google claims to be dedicated to user privacy, its reluctance to allow Americans to remove ‘irrelevant’ search results is “unfair and deceptive.”

I am not sure how quickly the various political bodies will move to make being forgotten a real thing. My hunch is that it will become an issue with legs. Down the road, the third party editing is likely to be required. The First Amendment is a hurdle, but when it comes times to fund a campaign or deal with winning an election, there may be some flexibility in third party editing’s appeal.

From my point of view, an index is an index. I have seen some frisky analyses of my blog articles and my for fee essays. I am not sure I want criticism of my work to be forgotten. Without an editorial policy, third party, ad hoc deletion of index pointers distorts the results as much, if not more, than results skewed by advertisers’ personal charm.

How about an editorial policy and then the application of that policy so that results are within applicable guidelines and representative of the information available on the public Internet?

Wow, that sounds old fashioned. The notion of an editorial policy is often confused with information governance. Nope. Editorial policies inform the database user of the rules of the game and what is included and excluded from an online service.

I like dinosaurs too. Like a cloned brontosaurus, is it time to clone the notion of editorial policies for corpus indices?

Stephen E Arnold, July 8, 2015

Compound Search Processing Repositioned at ConceptSearching

July 2, 2015

The article titled Metadata Matters; What’s The One Piece of Technology Microsoft Doesn’t Provide On-Premises Or in the Cloud? on ConceptSearching re-introduces Compound Search Processing, ConceptSearching’s main offering. Compound Search Processing is a technology achieved in 2003 that can identify multi-word concepts, and the relationships between words. Compound Search Processing is being repositioned, with Concept Searching apparently chasing Sharepoint Sales. The article states,

“The missing piece of technology that Microsoft and every other vendor doesn’t provide is compound term processing, auto-classification, and taxonomy that can be natively integrated with the Term Store. Take advantage of our technologies and gain business advantages and a quantifiable ROI…

Microsoft is offering free content migration for customers moving to Office 365…If your content is mismanaged, unorganized, has no value now, contains security information, or is an undeclared record, it all gets moved to your brand new shiny Office 365.”

The angle for Concept Searching is metadata and indexing, and they are quick to remind potential customers that “search is driven by metadata.” The offerings of ConceptSearching comes with the promise that it is the only platform that will work with all versions of Sharepoint while delivering their enterprise metadata repository. For more information on the technology, see the new white paper on Compoud Term Processing.
Chelsea Kerwin, July 2, 2014

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

CSC Attracts Buyer And Fraud Penalties

July 1, 2015

According to the Reuters article “Exclusive: CACI, Booz Allen, Leidos Eyes CSC’s Government Unit-Sources,” CACI International, Leidos Holdings, and Booz Allen Hamilton Holdings

have expressed interest in Computer Sciences Corp’s public sector division.  There are not a lot of details about the possible transaction as it is still in the early stages, so everything is still hush-hush.

The possible acquisition came after the news that CSC will split into two divisions: one that serves US public sector clients and the other dedicated to global commercial and non-government clients.  CSC has an estimated $4.1 billion in revenues and worth $9.6 billion, but CACI International, Leidos Holdings, and Booz Allen Hamilton might reconsider the sale or getting the price lowered after hearing this news: “Computer Sciences (CSC) To Pay $190M Penalty; SEC Charges Company And Former Executives With Accounting Fraud” from Street Insider.  The Securities and Exchange Commission are charging CSC and former executives with a $190 million penalty for hiding financial information and problems resulting from the contract they had with their biggest client.  CSC and the executives, of course, are contesting the charges.

“The SEC alleges that CSC’s accounting and disclosure fraud began after the company learned it would lose money on the NHS contract because it was unable to meet certain deadlines. To avoid the large hit to its earnings that CSC was required to record, Sutcliffe allegedly added items to CSC’s accounting models that artificially increased its profits but had no basis in reality. CSC, with Laphen’s approval, then continued to avoid the financial impact of its delays by basing its models on contract amendments it was proposing to the NHS rather than the actual contract. In reality, NHS officials repeatedly rejected CSC’s requests that the NHS pay the company higher prices for less work. By basing its models on the flailing proposals, CSC artificially avoided recording significant reductions in its earnings in 2010 and 2011.”

Oh boy!  Is it a wise decision to buy a company that has a history of stealing money and hiding information?  If the company’s root products and services are decent, the buyers might get it for a cheap price and recondition the company.  Or it could lead to another disaster like HP and Autonomy.

Whitney Grace, July 1, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Tumblr Has a GIF For You

June 30, 2015

Facebook recently enabled users to post GIF images on the social media platform.  Reddit was in an uproar over the new GIF and celebrated by posting random moving images from celebrities making weird faces to the quintessential cute kitten.  GIFs are an Internet phenomenon and are used by people to express their moods, opinions, or share their fandom.  Another popular social medium platform, Tumblr, the microblogging site used to share photos, videos, quotes, and more, has added a GIF search, says PCMag in “Tumblr Adds New GIF Search Capabilities.”

The main point of Tumblr is the ability share content either a user creates or someone else creates.  A user’s Tumblr page is a personal reflection of themselves and GIFs are one of the ultimate content pieces to share.  Tumblr’s new search option for GIFs is very simple: a user picks the + button, clicks the GIF button, and then search for the GIF that suits your mood.  A big thing on Tumblr is citing who created a piece and the new search option has that covered:

“Pick the GIF you want and it slinks right in, properly credited and everything,” the company said. “Whoever originally posted the GIF will be notified accordingly. On their dashboard, on their phone, all the regular places notifications go.”

GIFs are random bits of fun that litter the Internet and quickly achieve meme status.  They are also easy to make, which appeals to people with vey little graphic background.  They can make something creative and fun without much effort and now the can be easily found and shared on Tumblr.

Whitney Grace, June 30, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Digital Reasoning a Self-Described Cognitive Computing Company

June 26, 2015

The article titled Spy Tools Come to the Cloud on Enterprise Tech shows how Amazon’s work with analytics companies on behalf of the government have realized platforms like “GovCloud”, with increased security. The presumed reason for such platforms being the gathering of intelligence and threat analysis on the big data scale. The article explains,

“The Digital Reasoning cognitive computing tool is designed to generate “knowledge graphs of connected objects” gleaned from structured and unstructured data. These “nodes” (profiles of persons or things of interest) and “edges” (the relationships between them) are graphed, “and then being able to take this and put it into time and space,” explained Bill DiPietro, vice president of product management at Digital Reasoning. The partners noted that the elastic computing capability… is allowing customers to bring together much larger datasets.”

For former CIA staff officer DiPietro it logically follows that bigger questions can be answered by the data with tools like the AWS GovCloud and subsequent Hadoop ecosystems. He cites the ability to quickly spotlight and identify someone on a watch list out of the haystack of people as the challenge set to overcome. They call it “cluster on demand,” the process that allows them to manage and bring together data.

Chelsea Kerwin, June 26,  2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Twitter Gets a Search Facelift

June 25, 2015

Twitter has been experimenting with improving its search results and according to TechCrunch the upgrade comes via a new search results interface: “Twitter’s New Search Results Interface Expands To All Users.”  The new search results interface is the one of the largest updates Twitter has made in 2015.  It is supposed to increase the ease with a cleaner look and better filtering options.  Users will now be able to filter search results by live tweets, photos, videos, news, accounts, and more.

Twitter made the update to help people better understand how to use the message service and to take a more active approach to using it, rather than passively reading other peoples tweets.  The update is specifically targeted at new Twitter users.

The tweaked search interface will return tweets related to the search phrase or keyword, but that does not mean that the most popular tweets are returned:

“In some cases, the top search result isn’t necessarily the one with the higher metrics associated with it – but one that better matches what Twitter believes to be the searcher’s “intent.” For example, a search for “Steve Jobs” first displays a heavily-retweeted article about the movie’s trailer, but a search for “Mad Men” instead first displays a more relevant tweet ahead of the heavily-favorited “Mad Men” mention by singer Lorde.”

The new interface proves to be simpler and better list trends, related users, and news.  It does take a little while to finesse Twitter, which is a daunting task to new users.  Twitter is not the most popular social network these day and it’s using these updates to increase its appeal.

Whitney Grace, June 25, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

« Previous PageNext Page »

  • Archives

  • Recent Posts

  • Meta