Infonomics and the Big Data Market Publishers Need to Consider

March 22, 2016

The article on Beyond the Book titled Data Not Content Is Now Publishers’ Product floats a new buzzword in its discussion of the future of information: infonomics, or the study of creation and consumption of information. The article compares information to petroleum as the resource that will cause quite a stir in this century. Grace Hong, Vice-President of Strategic Markets & Development for Wolters Kluwer’s Tax & Accounting, weighs in,

“When it comes to big data – and especially when we think about organizations like traditional publishing organizations – data in and of itself is not valuable.  It’s really about the insights and the problems that you’re able to solve,”  Hong tells CCC’s Chris Kenneally. “From a product standpoint and from a customer standpoint, it’s about asking the right questions and then really deeply understanding how this information can provide value to the customer, not only just mining the data that currently exists.”

Hong points out that the data itself is useless unless it has been produced correctly. That means asking the right questions and using the best technology available to find meaning in the massive collections of information possible to collect. Hong suggests that it is time for publishers to seize on the market created by Big Data.

 

Chelsea Kerwin, March 22, 2016

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Search Dually Conceals and Retrieves for an Audience

March 8, 2016

There are many ways to trace a digital footprint, but Google is expanding European users’ ability to cover their tracks. An article entitled, In Europe, Google will now remove ‘right to be forgotten’ search results from all its sites, from The Verge tells us the story. Basically, European users who request links to be removed protect those links from being crawled by Google.fr in addition to all their other homepages. The write-up explains,

“So, for instance, if someone in France had previously requested that a link be hidden from search results, Google would just remove it from its European homepages, including google.fr. But a savvy searcher could have just used google.com to dig up all those hidden results. Now, however, the company will scrub its US homepage results, too, but only for European users. The company didn’t provide specifics on how it’ll detect that a user is in Europe, but it’s likely going off IP addresses, so in theory, someone could use a VPN to subvert those results.”

As the article mentions, European privacy regulators are happy about this but would still prefer contested links not appear, even if the searcher is in the U.S. or elsewhere. Between the existence of the Dark Web and the “right to be forgotten” protections, more and more links are hidden making search increasingly difficult.

 

Megan Feil, March 8, 2016

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Woman Fights Google and Wins

January 21, 2016

Google is one of those big corporations that if you have a problem with it, you might as well let it go.  Google is powerful, respected, and has (we suspect) a very good legal department.  There are problems with Google, such as the “right to be forgotten” and Australian citizens have a big bone to pick with the search engine.  Australian News reports that “SA Court Orders Google Pay Dr. Janice Duffy $115,000 Damages For Defamatory Search Results.”

Duffy filed a lawsuit against Google for displaying her name along with false and defamatory content within its search results.  Google claimed no responsibility for the actual content, as it was not the publisher.  The Australian Supreme Court felt differently:

“In October, the court rejected Google’s arguments and found it had defamed Dr Duffy due to the way the company’s patented algorithm operated.  Justice Malcolm Blue found the search results either published, republished or directed users toward comments harmful to her reputation.  On Wednesday, Justice Blue awarded Dr Duffy damages of $100,000 and a $15,000 lump sum to cover interest.”

Duffy was not the only one who was upset with Google.  Other Australians filed their own complaints, including Michael Trkulja with a claim search results linked him to crime and Shane Radbone sued to learn the identities of bloggers who wrote negative comments.

It does not seem that Google should be held accountable, but technically they are not responsible for the content.  However, Google’s algorithms are wired to bring up the most popular and in-depth results.  Should they develop a filter that measures negative and harmful information or is it too subjective?

 

Whitney Grace, January 21, 2016
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Xendo, Can Do

October 23, 2015

While it would be lovely to access and find all important documents, emails, and Web sites within a couple clicks, users usually have to access several programs or individual files to locate their information.  Stark Industries wanted users to have the power of Google search engine without compromising their personal security.  Xendo is a private, personal search engine that connects with various services, including email servers, social media account, clouds, newsfeeds, and more.

Once all the desired user accounts are connected to Xendo, the search engine indexes all the files within the services.  The index is encrypted, so it securely processes them.  After the indexing is finished, Xendo will search through all the files and return search results displaying the content and service types related to inputted keywords.  Xendo promises that:

“After your initial index is built, Xendo automatically keeps it up-to-date by adding, removing and updating content as it changes. Xendo automatically updates your index to reflect role and permission changes in each of your connected services. Xendo is hosted in some of the most secure data-centers in the world and uses multiple layers of security to ensure your data is secured in transit and at rest, like it’s in a bank vault.”

Basic Xendo search is free for individual users with payments required for upgrades.  The basic search offers deep search, unlimited access, and unlimited content, while the other plans offer more search options based on subscription.  Xendo can be deployed for enterprise systems, but it requires a personalized quote.

Whitney Grace, October 23, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Reclaiming Academic Publishing

October 21, 2015

Researchers and writers are at the mercy of academic publishers who control the venues to print their work, select the content of their work, and often control the funds behind their research.  Even worse is that academic research is locked behind database walls that require a subscription well beyond the price range of a researcher not associated with a university or research institute.  One researcher was fed up enough with academic publishers that he decided to return publishing and distributing work back to the common people, says Nature in “Leading Mathematician Launches arXiv ‘Overlay’ Journal.”

The new mathematics journal Discrete Analysis peer reviews and publishes papers free of charge on the preprint server arXiv.  Timothy Gowers started the journal to avoid the commercial pressures that often distort scientific literature.

“ ‘Part of the motivation for starting the journal is, of course, to challenge existing models of academic publishing and to contribute in a small way to creating an alternative and much cheaper system,’ he explained in a 10 September blog post announcing the journal. ‘If you trust authors to do their own typesetting and copy-editing to a satisfactory standard, with the help of suggestions from referees, then the cost of running a mathematics journal can be at least two orders of magnitude lower than the cost incurred by traditional publishers.’ ”

Some funds are required to keep Discrete Analysis running, costs are ten dollars per submitted papers to pay for software that manages peer review and journal Web site and arXiv requires an additional ten dollars a month to keep running.

Gowers hopes to extend the journal model to other scientific fields and he believes it will work, especially for fields that only require text.  The biggest problem is persuading other academics to adopt the model, but things move slowly in academia so it will probably be years before it becomes widespread.

Whitney Grace, October 21, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

The Tweet Gross Domestic Product Tool

October 16, 2015

Twitter can be used to figure out your personal income.  Twitter was not designed to be a tool to tally a person’s financial wealth, instead it is a communication tool based on a one hundred forty character messages to generate for small, concise delivery.  Twitter can be used to chat with friends, stars, business executives, etc, follow news trends, and even advertise products by sent to a tailored audience.  According to Red Orbit in the article “People Can Guess Your Income Based On Your Tweets,” Twitter has another application.

Other research done on Twitter has revealed that your age, location, political preferences, and disposition to insomnia, but your tweet history also reveals your income.  Apparently, if you tweet less, you make more money.  The controls and variables for the experiment were discussed, including that 5,191 Twitter accounts with over ten million tweets were analyzed and accounts with a user’s identifiable profession were used.

Users with a high follower and following ratio had the most income and they tended to post the least.  Posting throughout the day and cursing indicated a user with a lower income.  The content of tweets also displayed a plethora of “wealth” information:

“It isn’t just the topics of your tweets that’s giving you away either. Researchers found that “users with higher income post less emotional (positive and negative) but more neutral content, exhibiting more anger and fear, but less surprise, sadness and disgust.” It was also apparent that those who swore more frequently in their tweets had lower income.”

Twitter uses the information to tailor ads for users, if you share neutral posts get targeted ads advertising expensive items, while the cursers get less expensive ad campaigns.  The study also proves that it is important to monitor your Twitter profile, so you are posting the best side of yourself rather than shooting yourself in the foot.

Whitney Grace, October 16, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Politwoops Window Now Blackened

September 17, 2015

Why is Twitter helping politician around the world cover their tracks? The Bangkok Post reports, “Website that Saves Politicians’ Deleted Tweets Suspended.” Founded by the Open State Foundation as tool for transparency, Politwoops had made an agreement with Twitter in 2012 to use its API to post tweets that politicians (or their advisors) thought better of in retrospect. While Twitter reasons that any of their users should be able to take back tweets, the Open Foundation director asserts that public statements by public officials should remain part of the public record. The article states:

“Since being formed at a so-called hackathon five years ago, the website that is a useful tool for journalists and a frequent source of embarrassment for politicians, has spread to 30 countries from Egypt to the Vatican, as well as the European Parliament. It started operating in the US in 2012 thanks to the Sunlight Foundation, which fights for transparency in politics. Diplotwoops which screens deleted messages by diplomats and embassies around the world was set up in 2014. Twitter was not immediately available for comment, but the Open Foundation said it was told the social media giant decided to suspend access to Politwoops ‘following thoughtful internal deliberation and close consideration of a number of factors that doesn’t distinguish between users.’”

Um, except that one user is not like another. The public has a vested interest in knowing where elected officials stand, and it is tough to search when the content is no longer available. I wonder just what prompted Twitter’s “thoughtful internal deliberation.”

Cynthia Murrell, September 17, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Dark Web Drug Trade Unfazed by Law Enforcement Crackdowns

September 3, 2015

When Silk Road was taken down in 2013, the Dark Web took a big hit, but it was only a few months before black marketers found alternate means to sell their wares, including illegal drugs.  The Dark Web provides an anonymous and often secure means to purchase everything from heroin to prescription narcotics with, apparently, few worries about the threat of prosecution.  Wired explains that “Crackdowns Haven’t Stopped The Dark Web’s $100M Yearly Drug Sale,” proving that if there is a demand, the Internet will provide a means for illegal sales.

In an effort to determine if the Dark Web have grown to declined, Carnegie Mellon researchers Nicolas Cristin and Kyle Soska studied thirty-five Dark Web markets from 2013 to January 2015.  They discovered that the Dark Web markets are no longer explosively growing, but the market has remained stable fluctuating from $100 million to $180 million a year.

The researchers concluded that the Dark Web market is able to survive any “economic” shifts, including law enforcement crackdowns:

“More surprising, perhaps, is that the Dark Web economy roughly maintains that sales volume even after major disasters like thefts, scams, takedowns, and arrests. According to the Carnegie Mellon data, the market quickly recovered after the Silk Road 2 market lost millions of dollars of users’ bitcoins in an apparent hack or theft. Even law enforcement operations that remove entire marketplaces, as in last year’s purge of half a dozen sites in the Europol/FBI investigation known as Operation Onymous, haven’t dropped the market under $100 million in sales per year.”

Cristin and Soska’s study is the most comprehensive to measure the size and trajectory of the Dark Web’s drug market.  Their study ended prematurely, because two Web sites grew so big that the researchers’ software wasn’t able to track the content.  Their study showed that most Dark Web vendors are using more encryption tools, they make profits less $1000, and they are mostly selling MDMA and marijuana.

Soska and Cristin also argue that the Dark Web drug trade decreases violence in the retail drug trade, i.e. it keeps the transactions digital than having there be more violence on the streets.  They urge law enforcement officials to rethink shutting down the Dark Web markets, because it does not seem to have any effect.

Whitney Grace, September 3, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Suggestions for Developers to Improve Functionality for Search

September 2, 2015

The article on SiteCrafting titled Maxxcat Pro Tips lays out some guidelines for improved functionality when it comes deep search. Limiting your Crawls is the first suggestion. Since all links are not created equally, it is wise to avoid runaway crawls on links where there will always be a “Next” button. The article suggests hand-selecting the links you want to use. The second tip is Specify Your Snippets. The article explains,

“When MaxxCAT returns search results, each result comes with four pieces of information: url, title, meta, and snippet (a preview of some of the text found at the link). By default, MaxxCAT formulates a snippet by parsing the document, extracting content, and assembling a snippet out of that content. This works well for binary documents… but for webpages you wanted to trim out the content that is repeated on every page (e.g. navigation…) so search results are as accurate as possible.”

The third suggestion is to Implement Meta-Tag Filtering. Each suggestion is followed up with step-by-step instructions. These handy tips come from a partnering between Sitecrafting is a web design company founded in 1995 by Brian Forth. Maxxcat is a company acknowledged for its achievements in high performance search since 2007.

Chelsea Kerwin, September 2, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Sensible Advice on Content Marketing

August 21, 2015

Here’s a post on structured-content marketing that is refreshingly free of semantic search baloney. Tatiana Tilearcio at Synthesio shares what she learned from a seminar in, “Four Insights from a Content Marketing Crash Course.” The symposium, scheduled to be repeated in October in Connecticut, was presented by content-strategy outfit Content Boost. Tilearcio’s first takeaway promotes a firm foundation; she writes:

“Get Organized And Understand Your Goals Before You Create Your Content Marketing Plan.

Before you sit down to put together your strategic plan, you have to know the answer to the question ‘what’s the purpose for your content marketing, and what will it do to your brand?’ To do this, you need to first create a dream wish-list of what you would like to see for your brand. Next, you need to address how you want to go about enhancing your brand’s content marketing efforts and what your budget is. When creating a content marketing plan, or any marketing plan, a budget is essential. Without a proper budget of what your plan will cost, your ideas will never come to fruition. If you have identified all of this, then you are already well on your way to understanding what your campaign strategy is.”

The article also discusses blending efforts in blogging, social media, and email; co-sourcing content; ensuring users find value in gated assets; repurposing content; and the importance of strong titles. See the post for more details on each of these points. Based in Norwalk, Connecticut, Content Boost is part of the Technology Marketing Corporation, aka TMCnet.

Cynthia Murrell, August 21, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

« Previous PageNext Page »

  • Archives

  • Recent Posts

  • Meta