Are you frustrated by growing data requirements and interested in how non-relational databases could help? Curious about where and how NoSQL systems are being deployed? Want to build a "real" Highly Scalable System? Well if you answered yes to any of these questions you're not the only one who wants to know more. For this very reason we've decided to team up with some of the best NoSQL experts and bring to you the NoSQL Roadshow Amsterdam. It's been a great success in Stockholm, Copenhagen, Zurich & Basel and hence the roadtrip continues in Amsterdam and thereafter in London.
This informative and intensive 1 day session with 10 presentations is designed to give you an overview of the changing landscape around data management highlighted by concepts like "Big Data" and NoSQL databases, that are based on non-traditional, sometimes non-relational, other times very relational data models. The NoSQL Roadshow aims at IT professionals who are interested in faster and cheaper solutions to managing fast growing amounts of data.
After introducing the landscape and the business problems, you will learn how to attack these growing issues and hear first-hand how organisations were able to solve their modern data problems with innovative solutions like Neo4j, Riak, MongoDB, Cassandra, and many more.
NoSQL Roadshow Amsterdam is also a excellent opportunity for CIOs, CTOs, developers and NoSQL ambassadors to meet and discuss their own experiences with NoSQL - across industries and it's centrally located at the Pakhuis de Zwijger in the centre of the city.
We have limited capacity, so grab your seats today for Thursday 29th November and sign up for the early bird 50% discount rate of only 150 EUR.
We hope to see you there and help you get started with your big data decisions!
We’ve been providing Lucene & Solr consultancy and training services for quite a few years now and it’s always interesting to see how these two technologies are perceived by different companies and their technical people. More precisely, I find it interesting how little Solr users know about Lucene and more so, how unaware they are how important it is to to know about it. A quite reoccurring pattern we notice is that companies, looking for a cheap and good search solution, hear about Solr and decide to download and play around with it a bit. This is usually done within a context of a small PoC to eliminate initial investment risks. So one or two technical people are responsible for that, they download Solr distribution, and start following the Solr tutorial that is published on the Solr website. They realize that it’s quite easy to get things up and running using the examples Solr ships with and very quickly decide that this is the right way to go. So what the do next? They take their PoC codebase (including all Solr configurations) and slightly modify and extend them, just to support their real systems, and in no time, they get to the point were Solr can index all the data and then serve search requests. And that’s it... they roll out with it, and very often just put this in production. It is then often the case that after a couple of weeks we get a phone call from them asking for help. And why is that?
Examples are what they are - Just examples
I always argued that the examples that are bundled in the Solr distribution serve as a double edge sword. On one hand, they can be very useful just to showcase how Solr can work and provide good reference to the different setups it can have. On the other hand, it gives this false sense of security that if the examples configuration are good enough for the examples, they’ll be good enough for the other systems in production as well. In reality, this is of course far from being the case. The examples are just what they are - examples. It’s most likely that they are far from anything you’d need to support your search requirements. Take the Solr schema for example, this is one of the most important configuration files in Solr which contributes many of the factors that will influence the search quality. Sure, there are certain field types which you probably can always use (the primitive types), but when it comes to text fields and text analysis process - this is something you need to look closer at and in most cases customize to your needs. Beyond that, it’s also important to understand how different fields behave in respect to the different search functionality you need. What roles (if at all) can a field play in the context of these functionalities. For some functionalities (e.g. free text search) you need the fields to be analyzed, for other (e.g. faceting) you don’t. You need to have a very clear idea of these search functionalities you want to support, and based on that, define what normal/dynamic/copy fields should be configured. The examples configurations don’t provide you this insight as they are targeting the dummy data and the examples functionality they are aimed to showcase - not yours! And it’s not just about the schema, the solrconfig.xml in the examples is also much too verbose than you actually need/want it to be. Far too many companies just use these example configurations in their production environment and I just find it a pity. Personally, I like to view these configuration files also serving as some sort of documentation for your search solution - but by keeping them in a mess, full of useless information and redundant configuration, they obviously cannot.
It’s Lucene - not Solr
One of the greater misconceptions with Solr is that it’s a product on its own and that reading the user manual (which is an overstatement for a semi-structured and messy collection of wiki pages), one can just set it up and put it in production. What people fail to realize is that Solr is essentially just a service wrapper around Lucene, and that the quality of the search solution you’re building, largely depends on it. Yeah, sure... Solr provide important additions on top of Lucene like caching and few enhanced query features (e.g. function queries and dismax query parser), but the bottom line, the most influential factors of the search quality lays deep down in the schema definition which essentially determines how Lucene will work under the hood. This obviously requires proper understanding of Lucene... there’s just no way around it! But honestly, I can’t really “blame” users for getting this wrong. If you look at the public (open and commercial) resources that companies are selling to the users, they actually promote this ignorance by presenting Solr as a “stands on its own” product. Books, public trainings, open documentations, all hardly discuss Lucene in detail and instead focus more on “how you get Solr to do X, Y, Z”. I find it quite a shame and actually quite misleading. You know what? I truly believe that the users are smart enough to understand - on their own - what parameters they should send Solr to enable faceting on a specific field.... common... these are just request parameters so let them figure these things out. Instead, I find it much more informative and important to explain to them how faceting actually works under the hood. This way they understand the impact of their actions and configurations and are not left disoriented in the dark once things don’t work as they’d hoped. For this reason actually, we designed our Solr training to incorporate a relatively large portion of Lucene introduction in it. And take it from me... our feedback clearly indicate that the users really appreciate it!
There you have it... let it sink in: when downloading Solr, you’re also downloading Lucene. When configuring Solr, you’re also configuring Lucene. And if there are issues with Solr, they are often related to Lucene as well. So to really know Solr, do yourself a favor, and start getting to know Lucene! And you don’t need to be a Java developer for that, it’s not the code itself that you need to master. How Lucene works internally, on a detailed yet conceptual level should be more than enough for most users.
JTeam is proud to announce another public Wicket training. This public Wicket Introduction course (in cooperation with jWeekend) is scheduled on May 27 and 28. For more information (or on-site training inquiries) see: Apache Wicket Training page.
Register before April 15 and get a 25% discount!
Also, if you register more than one person, you even get more discount.
Signup now for the 3 day official Introduction to Apache Solr training, now available in Amsterdam!
Introduction to Solr is a 3 day instructor-led hands-on in-classroom training course, written by the engineers who helped write the Lucene/Solr code and led by JTeam's certified trainers. The objective of this course is to provide you with real use cases and teach you how to apply Solr search engine technologies to business requirements. During the course you will learn to apply best practices developing scalable, high availability and high performance search applications. View the course description.
This course is taught in collaboration with Lucid Imagination, the first commercial entity offering official training and support on Apache Solr and Lucene. Upon successful completion of the course, participants receive certification from Lucid.
Stay tuned for upcoming Lucene training dates.