How to Use Watson

August 7, 2015

While there are many possibilities for cognitive computing, what makes an idea a reality is its feasibility and real life application.  The Platform explores “The Real Trouble With Cognitive Computing” and the troubles IBM had (has) trying to figure out what they are going to do with the supercomputer they made.  The article explains that before Watson became a Jeopardy celebrity, the IBM folks came up 8,000 potential experiments for Watson to do, but only 20 percent of them.

The range is small due to many factors, including bug testing, gauging progress with fuzzy outputs, playing around with algorithmic interactions, testing in isolation, and more.  This leads to the “messy” way to develop the experiments.  Ideally, developers would have a big knowledge model and be able to query it, but that option does not exist.  The messy way involves keeping data sources intact, natural language processing, machine learning, and knowledge representation, and then distributed on an infrastructure.

Here is another key point that makes clear sense:

“The big issue with the Watson development cycle too is that teams are not just solving problems for one particular area. Rather, they have to create generalizable applications, which means what might be good for healthcare, for instance, might not be a good fit—and in fact even be damaging to—an area like financial services. The push and pull and tradeoff of the development cycle is therefore always hindered by this—and is the key barrier for companies any smaller than an IBM, Google, Microsoft, and other giants.”

This is exactly correct!  Engineering is not the same as healthcare and it not all computer algorithms transfer over to different industries.  One thing to keep in mind is that you can apply different methods from other industries and come up with new methods or solutions.

Whitney Grace, August 7, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Thunderstone Rumbles About Webinator

August 6, 2015

There is nothing more frustrating than being unable to locate a specific piece of information on a Web site when you use its search function.  Search is supposed to be quick, accurate, and efficient.  Even if Google search is employed as a Web site’s search feature, it does not always yield the best results.  Thunderstone is a company that specializes in proprietary software application developed specifically for information management, search, retrieval, and filtering.

Thunderstone has a client list that includes, but not limited to, government agencies, Internet developer, corporations, and online service providers.  The company’s goal is to deliver “product-oriented R&D within the area of advanced information management and retrieval,” which translates to them wanting to help their clients found information very, very fast and as accurately as possible.  It is the premise of most information management companies.  On the company blog it was announced that, “Thunderstone Releases Webinator Web Index And Retrieval System Version 13.”  Webinator makes it easier to integrate high quality search into a Web site and it has several new appealing features:

  • “Query Autocomplete, guides your users to the search they want
  • HTML Highlighting, lets users see the results in the original HTML for better contextual information
  • Expanded XML/SOAP API allows integration of administrative interface”

We like the HTML highlighting that offers users the ability to backtrack and see a page’s original information source. It is very similar to old-fashioned research: go back to the original source to check a fact’s veracity.

Whitney Grace, August 6, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Humans Screw Up the Self-Driving Car Again

August 5, 2015

Google really, really wants its self-driving cars to be approved for consumer usage.  While the cars have been tested on actual roads, they have also been accompanied by car accidents.  The Inquirer posted the article, “Latest Self-Driving Car Crash Injures Three Google Employees” about how the public might not be ready for self-driving vehicles.  Google, not surprisingly, blames the crash on humans.

Google has been testing self-driving cars for over six years and there have been a total of fourteen accidents involving the vehicles.  The most recent accident is the first that resulted in injuries.  Three Google employees were using the self-driving vehicle during Mountain View, California rush hour traffic on July 1. When the accident occurred, each of the three employees were treated for whiplash.  Google says that its car was not at fault and a distracted driver was at caused the accident, which is also the reason for the other accidents.

While Google is upset, the accidents have not hindered their plans, they have motivated them to push forward.  Google explained that:

“ ‘The most recent collision, during the evening rush hour on 1 July, is a perfect example. The light was green, but traffic was backed up on the far side, so three cars, including ours, braked and came to a stop so as not to get stuck in the middle of the intersection.  After we’d stopped, a car slammed into the back of us at 17 mph? ?and it hadn’t braked at all.’ ”

Google continues to insist that human error and inattention are ample reason to allow self-driving cars on the road.  While it is hard to trust a machine with driving a weapon going 50 miles per hour, why do we trust people who have proven to be poor drivers with a license?

Whitney Grace, August 5, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Hire Watson As Your New Dietitian

August 4, 2015

IBM’s  supercomputer Watson is being “trained” in various fields, such as healthcare, app creation, customer service relations, and creating brand new recipes.  The applications for Watson are possibly endless.  The supercomputer is combining its “skills” from healthcare and recipes by trying its hand at nutrition.  Welltok invented the CaféWell Health Optimization Platform, a PaaS that creates individualized healthcare plans, and it implemented Watson’s big data capabilities to its Healthy Dining CaféWell personal concierge app.  eWeek explains that “Welltok Takes IBM Watson Out To Dinner,” so it can offer clients personalized restaurant menu choices.

” ‘Optimal nutrition is one of the most significant factors in preventing and reversing the majority of our nation’s health conditions, like diabetes, overweight and obesity, heart disease and stroke and Alzheimer’s,’ said Anita Jones-Mueller, president of Healthy Dining, in a statement. ‘Since most Americans eat away from home an average of five times each week and it can be almost impossible to know what to order at restaurants to meet specific health needs, it is very important that wellness and condition management programs empower  smart dining out choices. We applaud Welltok’s leadership in providing a new dimension to healthy restaurant dining through its groundbreaking CaféWell Concierge app.’”

Restaurant menus are very vague when it comes to nutritional information.  When it comes to knowing if something is gluten-free, spicy, or a vegetarian option, the menu will state it, but all other information is missing.  In order to find a restaurant’s nutritional information, you have to hit the Internet and conduct research.  A new law passed will force restaurants to post calorie counts, but that will not include the amount of sugar, sodium, and other information.  People have been making poor eating choices, partially due to the lack of information, if they know what they are eating they can improve their health.  If Watson’s abilities can decrease the US’s waistline, it is for the better.  The bigger challenge would be to get people to use the information.

Whitney Grace, August 4, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

Bodleian Library Gets Image Search

August 3, 2015

There is a lot of free information on the Internet, but the veracity is always in question.  While libraries are still the gateway of knowledge, many of their rarer, more historic works are buried in archives.  These collections offer a wealth of information that is often very interesting.  The biggest problem is that libraries often lack the funds to scan archival collections and create a digital library.  Oxford University’s Bodleian Library, one of the oldest libraries in Europe, has the benefit of funds and an excellent collection to share with the world.

Digital Bodleian boasts over 115,179 images as of writing this article, stating that it is constantly updating the collection.  The online library takes a modern approach to how users interact with the images by taking tips from social media.  Not only can users browse and search the images randomly or in the pre-sorted collections, they can also create their own custom libraries and sharing the libraries with friends.

It is a bold move for a library, especially for one as renowned as Bodleian, to embrace a digital collection as well as offering a social media-like service.  In my experience, digital library collections are bogged down by copyright, incomplete indices or ontologies, and they lack images to perk a users’ interest.  Digital Bodleian is the opposite of many of its sister archives, but another thing I have noticed is that users are not too keen on joining a library social media site.  It means having to sign up for yet another service and also their friends probably aren’t on it.

Here is an idea, how about a historical social media site similar to Pinterest that pulls records from official library archives?  It would offer the ability to see the actual items, verify information, and even yield those clickbait top ten lists.

Whitney Grace, August 3, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Online Ads Discriminate

August 3, 2015

In our modern age, discrimination is supposed to be a thing of the past.  When it does appear, people take to the Internet to vent their rage and frustrations, eager to point out this illegal activity.  Online ads, however, lack human intelligence and are only as smart as their programmed algorithm.  Technology Review explains in “Probing The Dark Side of Google’s Ad-Targeting System” that Google’s ad service makes inaccurate decisions when it comes to gender and other personal information.

A research team at Carnegie Mellon University and the International Computer Science Institute built AdFisher, a tool to track targeted third party ads on Google.  AdFisher found that ads were discriminating against female users.  Google offers a transparency tool that allows users to select what types of ads appear on their browsers, but even if you use the tool it doesn’t stop some of your personal information from being used.

“What exactly caused those specific patterns is unclear, because Google’s ad-serving system is very complex. Google uses its data to target ads, but ad buyers can make some decisions about demographics of interest and can also use their own data sources on people’s online activity to do additional targeting for certain kinds of ads. Nor do the examples breach any specific privacy rules—although Google policy forbids targeting on the basis of “health conditions.” Still, says Anupam Datta, an associate professor at Carnegie Mellon University who helped develop AdFisher, they show the need for tools that uncover how online ad companies differentiate between people.”

The transparency tool only controls some of the ads and third parties can use their own tools to extract data.  Google stands by its transparency tool and even offers users the option to opt-out of ads.  Google is studying AdFisher’s results and seeing what the implications are.

The study shows that personal data spills out on the Internet every time we click a link or use a browser.  It is frightening how the data can be used and even hurtful if interpreted incorrectly by ads.  The bigger question is not how retailers and Google uses the data, but how do government agencies and other institutes plan to use it?

Whitney Grace, August 3, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Bing Is Very Important, I Mean VERY Important

July 31, 2015

The online magazine eWeek published, “What The Bing Search Engine Brings To Microsoft’s Web Strategy” and it explains how Bing spurs a lot of debate:

“Some who don’t like the direction in which Google is going say that Bing is the search engine they prefer, especially since Microsoft has honed Bing’s ability to deliver relevant results. Others, however, look at Bing as one of many products from Microsoft, which is still seen as the “Evil Empire” in some quarters and a search platform that’s incapable of delivering the results that compare favorably with Google. Bing, introduced six years ago in 2009, is still a remarkably controversial product in Microsoft’s lineup. But it’s one that plays an important role in so many of the company’s Internet services.”

Microsoft is ramping up Bing to become a valuable part of its software services, it continues its partnership with Yahoo and Apple, and it will also power AOL’s web advertising and search.  Bing is becoming a more respected search engine, but what does it have to offer?

Bing has many features it is using to entice people to stop using Google.  When searching a person’s name, search results display a bio of the person (only if they are affluent, however).  Bing has a loyalty program, seriously, called Bing Rewards, the more you search on Bing it rewards points that are redeemable for gift cards, movie rentals, and other items.

Bing is already a big component in Microsoft software, including Windows 10 and Office 365.  It serves as the backbone for not only a system search, but searching the entire Internet.  Think Apple’s Spotlight, except for Windows.  It also supports a bevy of useful applications and do not forget about Cortana, which is Microsoft’s answer to Siri.

Bing is very important to Microsoft because of the ad revenue.  It is just a guess, but you can always ask Cortana for the answer.

Whitney Grace, July 31, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

 

Finnish Content Discovery Case Study

July 31, 2015

There are many services that offer companies the ability to increase their content discover.  One of these services is Leiki, which offers intelligent user profiling, context-based intelligence, and semantic SaaS solutions.  Rather than having humans adapt their content to get to the top of search engine results, the machine is altered to fit a human’s needs.  Leiki pushes relevant content to a user’s search query.  Leiki released a recent, “Case Study: Lieki Smart Services Increase Customer Flow Significantly At Alma Media.”

Alma Media is one of the largest media companies in Finland, owning many well-known Finnish brands.  These include Finland’s most popular Web site, classified ads, and a tabloid newspaper.  Alma Media employed two of Leiki’s services to grow its traffic:

“Leiki’s Smart Services are adept at understanding textual content across various content types: articles, video, images, classifieds, etc. Each content item is analyzed with our semantic engine Leiki Focus to create a very detailed “fingerprint” or content profile of topics associated with the content.

SmartContext is the market leading service for contextual content recommendations. It’s uniquely able to recommend content across content types and sites and does this by finding related content using the meaning of content – not keyword frequency.

SmartPersonal stands for behavioral content recommendations. As it also uses Leiki’s unique analysis of the meaning in content, it can recommend content from any other site and content type based on usage of one site.”

The case study runs down how Leiki’s services improved traffic and encouraged more users to consume its content. Leiki’s main selling point in the cast study is that offers users personal recommendations based on content they clicked on Alma Media Web sites.  Leiki wants to be a part of developing Web 3.0 and the research shows that personalization is the way for it to go.

Whitney Grace, July 31, 2015
Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

Organizations Should Consider Office 365 Utilization

July 30, 2015

Office 365 has been a bit contentious within the community. While Microsoft touts it as a solution that gives users more of the social and mobile components they were wishing for, it has not been widely adopted. IT Web gives some reasons to consider the upgrade in its article, “Why You Should Migrate SharePoint to Office 365.”

The article says:

“Although SharePoint as a technology has matured a great deal over the years, I still see many businesses struggling with issues related to on-premises SharePoint, says Simon Hepburn, director of bSOLVe . . . You may be thinking: ‘Are things really that different using SharePoint on Office 365?’ Office 365 is constantly evolving and as I will explain, this evolution brings with it opportunities that your business should seriously consider exploring.’”

Of course the irony is that with the new SharePoint 2016 upgrade, Microsoft is giving users a promise to stand behind on-premise installations, but they are continuing to integrate and promote the Office 365 components. Only time and feedback will dictate the continued direction of the enterprise solution. In the meantime, stay tuned to Stephen E. Arnold and his Web service, ArnoldIT.com. Arnold is a longtime leader in search and his dedicated SharePoint feed is a one-stop-shop for all the latest news, tips, and tricks.

Emily Rae Aldridge, July 30, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

 

Whither Unix Data

July 30, 2015

For anyone using open-source Unix to work with data, IT World has a few tips for you in “The Best Tools and Techniques for Finding Data on Unix Systems.” In her regular column, “Unix as a Second Language,” writer Sandra Henry-Stocker explains:

“Sometimes looking for information on a Unix system is like looking for needles in haystacks. Even important messages can be difficult to notice when they’re buried in huge piles of text. And so many of us are dealing with ‘big data’ these days — log files that are multiple gigabytes in size and huge record collections in any form that might be mined for business intelligence. Fortunately, there are only two times when you need to dig through piles of data to get your job done — when you know what you’re looking for and when you don’t. 😉 The best tools and techniques will depend on which of these two situations you’re facing.”

When you know just what to search for, Henry-Stocker suggests the “grep” command. She supplies a few variations, complete with a poetic example. Sometimes, like when tracking errors, you’re not sure what you will find but do know where to look. In those cases, she suggests using the “sed” command. For both approaches, Henry-Stocker supplies example code and troubleshooting tips. See the article for the juicy details.

Cynthia Murrell, July 30, 2015

Sponsored by ArnoldIT.com, publisher of the CyberOSINT monograph

 

« Previous PageNext Page »

  • Archives

  • Recent Posts

  • Meta