Microsoft Cloud Authors: Janakiram MSV, Yeshim Deniz, David H Deans, Andreas Grabner, Stackify Blog

Related Topics: Microsoft Cloud

Microsoft Cloud: Blog Feed Post

Search in SharePoint 2013 Preview

Search is everywhere in SharePoint now so it is important to understand how it has changed

As I start writing this post, I know this is going to be one of those posts that covers so much that it is going to hard to cover it all.  The point of this post is to give you a high level idea of everything new and changed in Search with SharePoint 2013 Preview.  Search is everywhere in SharePoint now so it is important to understand how it has changed.  For example, take a look at this document library utilizing the power of Search.


Aside from cool stuff like drag and drop, document libraries directly leverage the search engine to allow users to filter documents easily.  Gone are the days of relying on CAML for simple document library searches.  You’ll find out why later in this post on why you can rely on search for those uses as well.

In SharePoint 2013 Preview, Search was essentially rewritten from the ground up.  You’ll see familiar concepts like managed properties as well as how FAST Search for SharePoint morphed into this new product.  The concept of FAST Search for SharePoint servers is gone.  The components from FS4SP have made their way directly into SharePoint and we don’t need separate dedicated servers for it (necessarily).  If you’re familiar with FAST ESP, you will see some familiar components from there too.  Don’t worry, we aren’t starting over from scratch, but you will see some exciting new things that make search such a powerful feature of SharePoint.

Today’s post will provide a high level of many of the new concepts in search. It will serve as a springboard for a series of detailed articles about the individual components of Search that I will post in the coming weeks.  We’ll cover Search in the following areas: topology, crawling, querying, user interface, API, and SharePoint Online.


The underlying search topology has changed quite a bit.  However, most of it is based upon concepts you may have seen before from FAST Search for SharePoint.  Components can be scaled out to multiple servers as needed.  These changes can be done through the Search Service Application or through PowerShell.  When making changes to the topology, you don’t change the active topology, you clone the original, make changes, and then change the active topology.  This section is pretty technical so feel free to skip it if you are only interested in the cool stuff like querying and the user interface. :)


The first thing to cover is how the topology changed.  Many components got new (but similar) names and they correspond to FAST Search for SharePoint components.  The Search components are hosted on your SharePoint application servers using a Search Service Application.  Components can be scaled to multiple servers for performance and redundancy.  The components that make up search are:

Crawl Component

The crawl component crawls the actual data from a variety of sources such as SharePoint, File Shares, User Profiles, and Databases using BCS.

Content Processing Component

This component processes crawled items and feeds them to the index component.  This is where document parsing occurs as well as IFilters exist.  A generic IFIilter will cover most of your needs.  It also is responsible for language detection and entity extraction (both of which are features from FS4SP).  It also produces the phonetic name variations for people search.

Index Component

If you are familiar with FAST Search for SharePoint, the index component will look pretty similar.  The Index Component is used in both feeding and query processes.  It takes items from the content processing component and writes them to the index.  It also receives queries from the query processing component and returns result sets.  The Index architecture is based off the rows and columns concepts in FS4SP.  Index Replicas (rows) provide a level of redundancy with groups of servers.  Index Partitions (columns) allow you to split the index between servers.

Analytics Component

The Analytics Component analyzes crawled items and how users interact with Search Results.  It truly is a part of what makes search “learn” and provide better search results to the user.

Query Processing Component

This component performs linguistic processing at query time such as work breaking, stemming, spell checking, and the thesaurus.  When the query comes in, it completes it’s processing and passes it to the query component.

Search Administration Component

The administration component stores the various information about search that you configure through the user interface in the Search Service Application.  It also manages topology changes.

Search Processes

The search processes look a bit different than what you may be used to.  The Host Controller is a Windows service that manages various processes called NodeRunners.  When you first install SharePoint 2013 Preview and you wonder where all your memory is, you’ll see multiple NodeRunner.exe processes at the top of the list.  Each NodeRunner.exe hosts one of the various components above.  Looking at the task manager, it is not obvious which process is running which component.  There is a PowerShell script that will tell you, that I will post in the future.  Lastly, MSSearch.exe is the Windows Service that hosts the crawl component.


The configuration of crawling looks similar but there are lots of changes.  Many of the configuration changes you make aren’t limited to the Search Service Application any more.  Changes can be made at the site collection and some can even be made at the site.

Content Sources

Content Sources are configured in much the same way, but there are some changes.  The newest feature here is the concept of Continuous Crawling.  This crawls your content source continuously (every 15 minutes by default).  However, there is some magic that occurs now and new items can appear in the index within seconds.  This is something users have always wanted and I am really excited about it.  It also means that when the full crawl is executing, you can see changes to the index while it is still running.

Result Sources

Results Sources effectively combine Scopes and Federated Locations into one interface.  However, they added a ton of new features in how you can build the queries that make up the result source.  In a new instance, quite a few result sources are available out-of-the-box such as Local SharePoint Results, Popular, and Items Matching a Content Type.  Here is what it looks like.


Whereas SharePoint 2010, only had protocols for Local Search and OpenSearch 1.1, SharePoint 2013 Preview (as well as SharePoint Online Preview), support for Remote SharePoint servers and Exchange has been added.  Since SharePoint and People Search results are served by the same search index now, you can choose which type of results you want here too.


Scrolling down the page, you have the ability to use the new Query Builder to construct a query.  This new interface provides a lot of ways to create custom queries very easily.


The query builder lets you quickly construct a query even with dynamic values.


After you construct your query, you can click the TEST tab to see if the query works.  There is so much to cover with the Query Builder, it will gets it own post in the near future.


Document Parsing

A number of improvements have been made to document parsing as well.  New high performance IFilters exist for common Office document formats as well as images and PDF.  This means you won’t have to manually configure that IFilter any more.  However, the existing IFilter interface is available still in case there is anything you want to add.

Entity Extraction

In FS4SP, entity extraction was managed using a set of XML files.  In SharePoint 2013 Preview, this has been moved to the term store.  The out-of-the-box entity extraction will automatically extract company names out of documents.  You can use the term store to manage exclusions and inclusions.  Unlike FS4SP, it doesn’t look like you are able to add your own term sets for entity extraction, so this is unfortunate.


Schema Management

Managed Properties and Crawled Properties are now referred to as the Search Schema.  Managed Properties have a number of new parameters such as allowing for sorting and refining that we gained from FS4SP.


Site Collection administrators now also have the ability to make changes to the search schema at their level.  This allows you to delegate some of the search configuration to admins and let them override settings without affecting things globally.


Export and Import

Another one of my favorite features in Search is that we finally have the ability to export and import search configurations.  I’ve only been asking for this since 2007. :)  You can export your search configuration as XML and then import it later.  This can even be done at the site collection level.


Crawl Log Permissions

There are times where you may want to grant access to the crawl logs to non-administrator users.  The new setting allows you to grant other users access.



We’ll now look at some of the improvements when querying search.

Query Spelling Correction

Customizations to the Spelling Correction is now managed through the term store as well.  This allows you to customize the “Did you mean?” functionality.

Query Rules

The new query rules engine lets you tailor your query results in ways never before possible.  Each rule is composed of conditions and actions.  When the conditions are met, one or more actions are implemented.  Actions include things like promoting a result (similar to a best bet) and injecting a result block into the search results (basically changing the way search results look).  I think a screenshot of the out-of-the-box query rules actually explains it better than I do.


In this example, if it finds a person name in the Local SharePoint Results, it will promote a result block showing people that matched the result.  You can create your own query rules to really customize how search results look.  You can even put start and end dates on a query rule.

User Interface

The User Interface has pretty much been rewritten in its entirety.  It starts with the addition of the ResultScriptWebPart which retrieves and displays search results.  This replaces the CoreResultsWebPart from SharePoint 2010 and has a ton of new functionality.  Take a look at the search center in this example:


There is a lot to take in here.  In my example here, I had a number of PowerPoint presentations returned.  When I hover over one of them, it gives me a large visual preview of the slide deck that I can flip through.  On top of that, it picked up the key sections of the document and listed them under “Take a look inside”.  At the bottom of the preview there are useful links such as Follow (Social feature), Edit, View Library, and Send.  What’s cool about the document preview is that it lets you scroll through the entire document.  It even shows the animations in PowerPoint decks.  If you used this functionality at all before with FS4SP, you might have been hit with the fact that document previews didn’t work against documents sitting on a claims-enabled web application.  Note, that previews only work with claims authentication now (along with many other things).

There are a few remaining features to point out.  On the left, you see some of the new visual refiners that allow you to search by different modified dates.  The search box at the top also provides options to easily jump between documents, people, conversations, videos, and reports.

Result Types

In the above screenshot, you might have noticed that the PowerPoint results are formatted a certain way.  This is through the new Result Type feature that allows you to customize how a particular result looks based upon a condition.  As someone that customizes search, if there is anything to get excited about, this is it.  Result Types are comprised of Rules, Properties, and Display Templates.  The Rules define when the result type should be used (i.e.: Excel Documents, People, or Picture Library).  When one of these rules matches, that Result Type will be used to display the individual result in a unique way.  Properties refer to managed properties and these are what you will use in your display template to show the data from the result.  Here is what the Result Types page looks like in the Site Collection.


You can edit the built-in result types, but you can create new ones.  Before you define your result type, it is a good idea to create a new display template as you will have to select it from a list when creating the result type.  The Display Templates are .js files and are kept in the ~sitecollection/_catalogs/masterpage/Display Templates/Search folder.

I’ll post soon about how to configure result types, but take a look at another example.  This time I have a mixture of different document types.  Most result types look the same out of the box, but you can customize them heavily to meet your needs.


Search Refinement

The refiner web part has some added functionality.  It has the most of refinement included in SharePoint 2010.  However, for those of you who used to use FAST ESP, we now have faceted navigation.  This allows you to use refinement before you ever issue a search and is based upon data in the term store.  Think of how BestBuy.com uses FAST ESP to allow you to select TVs –> >50” –> Plasma, etc.  We can also leverage display templates to change how the refinement is rendered.  This makes use of the new Refinable attribute that we see on managed properties.  For example, that is how you see the data slider.


Query Suggestions

Query Suggestions have been improved largely through the use of the analytics component to analyze your personal search history.  It actually weights results based on links you have previously visited.  It also looks at the most frequent queries of all users to deliver better suggestion.  There are two types of query suggestions: what you see before issuing the query and what you see along with the results.  For the pre-query suggestions, you will get suggestions from your personal query log along with what other people have been searching for.  For the suggestions after you get results back, it returns matches that you have clicked on at least twice.  I’ll post more on this later after I have an environment that has been up long enough to capture some of this data.


The Search API has underwent a series of changes.  Two new interfaces are available, while one was removed and another deprecated.  If you write custom search code, you will want to pay attention to this section.

New Interfaces

The SharePoint REST API got some love in this release.  In terms of search, we have a whole new interface for querying using REST.  This is possible by calling the endpoint located at /_api/search.  You can specify any site collection or site in the URL, but typically you’ll just go with the web application root URL.  Specify any other URL prefix will get you the same results as well.  In one of my upcoming posts, we’ll go into some real examples of how to use this new endpoint.

Search also got some love in the Client OM.  That means you can now execute search queries using CSOM via JavaScript or .NET.  According to the MSDN post, you can do mobile development as well, but I’m unsure if that includes Windows Phone.  I am thinking it doesn’t but you can still use the REST API there.  I’ll confirm that as I begin writing my posts on the API.

Removed and Deprecated

The Search API has been expanded greatly and you have some great new options to use.  However, there are a few other changes you need to know about.  First, and foremost, the SQL Syntax has been removed.  I’ve been telling you for years to stop using it.  At SPC09, they said it was unofficially deprecated and now I can say straight up that it is gone now.  You also need to know that the Search web service (search.asmx) is now deprecated (but not gone) as well.  That means you need to stop using the web service as it won’t work some day.  If you need to remotely access Search, then use the Query CSOM or the Query REST Service which are much better and featured.

New Operators

If you have been using FAST Search for SharePoint, these three new operators will be nothing new to you.  However, if you were strictly running SharePoint 2010, they may be of interest.  Previously the XRANK operator was only available in FQL.  Now we can use it in regular keyword queries and it gives us the ability to dynamically adjust the rank of items.  The NEAR operator has been improved to include a configurable token distance (besides the default of 8) and a new ONEAR operator allows for ordered near functionality.  Most of these operators are pretty hardcore so most people probably will never use them but they are there if you need them.  You can also continue to use FQL if you prefer.

SharePoint Online

As Microsoft works to bring feature parity to SharePoint Online, the preview brings us a heap of new features in Search.  Whereas you could configure next to nothing in the previous iteration of SharePoint Online, you can do just about anything with Search now.  This comes from the new Search link inside tenant administration.  From the list you see below, you can do just about anything except configure Content Sources.  These are still handles automatically by SharePoint Online so we can’t change the frequency of crawls nor can we crawl other sources such as HTTP or BCS.  Take a look at the Search settings in Tenant administration to get a feel for what you can do.


Scrolling down, we can adjust a few more settings.


I’ve already talked about Export and Import.  The new Search Center Settings link allows you to set a global search center that will be used on each site.  The feature parity in Search with SharePoint Online Preview is impressive.  In fact, all of my screenshots for this article came from the cloud.


As I expected, this post has proved to be quite long.  I tried to be brief in each section so as I could cover as much as possible.  Anything in bold in this article will likely be a follow-up blog post, so stay tuned.  I expect to find information that needs to be updated or points that I left out. I’ll be posting updates to this post as necessary. Anyhow, I hope this post has proved useful in explaining what you need to know about Search in SharePoint 2013 Preview.

Follow me on twitter at @coreyroth.

Read the original blog entry...

More Stories By Corey Roth

Corey Roth, a SharePoint Server MVP, is a consultant at Hitachi Consulting specializing in SharePoint and Office 365 for clients in the energy sector. He has more than ten years of experience delivering solutions in the energy, travel, advertising and consumer electronics verticals.

Corey specializes in delivering ECM and search solutions to clients using SharePoint. Corey has always focused on rapid adoption of new Microsoft technologies including Visual Studio 2013, Office 365, and SharePoint.

He is a member of the .NET Mafia (www.dotnetmafia.com) where he blogs about the latest technology and SharePoint. He is dedicated to the community and speaks regularly at user groups and SharePoint Saturdays.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.

@ThingsExpo Stories
"Cloud Academy is an enterprise training platform for the cloud, specifically public clouds. We offer guided learning experiences on AWS, Azure, Google Cloud and all the surrounding methodologies and technologies that you need to know and your teams need to know in order to leverage the full benefits of the cloud," explained Alex Brower, VP of Marketing at Cloud Academy, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clar...
In his session at 21st Cloud Expo, Carl J. Levine, Senior Technical Evangelist for NS1, will objectively discuss how DNS is used to solve Digital Transformation challenges in large SaaS applications, CDNs, AdTech platforms, and other demanding use cases. Carl J. Levine is the Senior Technical Evangelist for NS1. A veteran of the Internet Infrastructure space, he has over a decade of experience with startups, networking protocols and Internet infrastructure, combined with the unique ability to it...
"IBM is really all in on blockchain. We take a look at sort of the history of blockchain ledger technologies. It started out with bitcoin, Ethereum, and IBM evaluated these particular blockchain technologies and found they were anonymous and permissionless and that many companies were looking for permissioned blockchain," stated René Bostic, Technical VP of the IBM Cloud Unit in North America, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventi...
Gemini is Yahoo’s native and search advertising platform. To ensure the quality of a complex distributed system that spans multiple products and components and across various desktop websites and mobile app and web experiences – both Yahoo owned and operated and third-party syndication (supply), with complex interaction with more than a billion users and numerous advertisers globally (demand) – it becomes imperative to automate a set of end-to-end tests 24x7 to detect bugs and regression. In th...
Widespread fragmentation is stalling the growth of the IIoT and making it difficult for partners to work together. The number of software platforms, apps, hardware and connectivity standards is creating paralysis among businesses that are afraid of being locked into a solution. EdgeX Foundry is unifying the community around a common IoT edge framework and an ecosystem of interoperable components.
"MobiDev is a software development company and we do complex, custom software development for everybody from entrepreneurs to large enterprises," explained Alan Winters, U.S. Head of Business Development at MobiDev, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Large industrial manufacturing organizations are adopting the agile principles of cloud software companies. The industrial manufacturing development process has not scaled over time. Now that design CAD teams are geographically distributed, centralizing their work is key. With large multi-gigabyte projects, outdated tools have stifled industrial team agility, time-to-market milestones, and impacted P&L stakeholders.
"Akvelon is a software development company and we also provide consultancy services to folks who are looking to scale or accelerate their engineering roadmaps," explained Jeremiah Mothersell, Marketing Manager at Akvelon, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"Space Monkey by Vivent Smart Home is a product that is a distributed cloud-based edge storage network. Vivent Smart Home, our parent company, is a smart home provider that places a lot of hard drives across homes in North America," explained JT Olds, Director of Engineering, and Brandon Crowfeather, Product Manager, at Vivint Smart Home, in this SYS-CON.tv interview at @ThingsExpo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Coca-Cola’s Google powered digital signage system lays the groundwork for a more valuable connection between Coke and its customers. Digital signs pair software with high-resolution displays so that a message can be changed instantly based on what the operator wants to communicate or sell. In their Day 3 Keynote at 21st Cloud Expo, Greg Chambers, Global Group Director, Digital Innovation, Coca-Cola, and Vidya Nagarajan, a Senior Product Manager at Google, discussed how from store operations and ...
"There's plenty of bandwidth out there but it's never in the right place. So what Cedexis does is uses data to work out the best pathways to get data from the origin to the person who wants to get it," explained Simon Jones, Evangelist and Head of Marketing at Cedexis, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that CrowdReviews.com has been named “Media Sponsor” of SYS-CON's 22nd International Cloud Expo, which will take place on June 5–7, 2018, at the Javits Center in New York City, NY. CrowdReviews.com is a transparent online platform for determining which products and services are the best based on the opinion of the crowd. The crowd consists of Internet users that have experienced products and services first-hand and have an interest in letting other potential buye...
SYS-CON Events announced today that Telecom Reseller has been named “Media Sponsor” of SYS-CON's 22nd International Cloud Expo, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Telecom Reseller reports on Unified Communications, UCaaS, BPaaS for enterprise and SMBs. They report extensively on both customer premises based solutions such as IP-PBX as well as cloud based and hosted platforms.
It is of utmost importance for the future success of WebRTC to ensure that interoperability is operational between web browsers and any WebRTC-compliant client. To be guaranteed as operational and effective, interoperability must be tested extensively by establishing WebRTC data and media connections between different web browsers running on different devices and operating systems. In his session at WebRTC Summit at @ThingsExpo, Dr. Alex Gouaillard, CEO and Founder of CoSMo Software, presented ...
WebRTC is great technology to build your own communication tools. It will be even more exciting experience it with advanced devices, such as a 360 Camera, 360 microphone, and a depth sensor camera. In his session at @ThingsExpo, Masashi Ganeko, a manager at INFOCOM Corporation, introduced two experimental projects from his team and what they learned from them. "Shotoku Tamago" uses the robot audition software HARK to track speakers in 360 video of a remote party. "Virtual Teleport" uses a multip...
A strange thing is happening along the way to the Internet of Things, namely far too many devices to work with and manage. It has become clear that we'll need much higher efficiency user experiences that can allow us to more easily and scalably work with the thousands of devices that will soon be in each of our lives. Enter the conversational interface revolution, combining bots we can literally talk with, gesture to, and even direct with our thoughts, with embedded artificial intelligence, whic...
SYS-CON Events announced today that Evatronix will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Evatronix SA offers comprehensive solutions in the design and implementation of electronic systems, in CAD / CAM deployment, and also is a designer and manufacturer of advanced 3D scanners for professional applications.
Leading companies, from the Global Fortune 500 to the smallest companies, are adopting hybrid cloud as the path to business advantage. Hybrid cloud depends on cloud services and on-premises infrastructure working in unison. Successful implementations require new levels of data mobility, enabled by an automated and seamless flow across on-premises and cloud resources. In his general session at 21st Cloud Expo, Greg Tevis, an IBM Storage Software Technical Strategist and Customer Solution Architec...
To get the most out of their data, successful companies are not focusing on queries and data lakes, they are actively integrating analytics into their operations with a data-first application development approach. Real-time adjustments to improve revenues, reduce costs, or mitigate risk rely on applications that minimize latency on a variety of data sources. In his session at @BigDataExpo, Jack Norris, Senior Vice President, Data and Applications at MapR Technologies, reviewed best practices to ...
An increasing number of companies are creating products that combine data with analytical capabilities. Running interactive queries on Big Data requires complex architectures to store and query data effectively, typically involving data streams, an choosing efficient file format/database and multiple independent systems that are tied together through custom-engineered pipelines. In his session at @BigDataExpo at @ThingsExpo, Tomer Levi, a senior software engineer at Intel’s Advanced Analytics gr...