We’ve all dreamed of going to bed one day and waking up the next with superpowers – stronger, faster and even perhaps with the ability to fly. Yesterday that is exactly what happened to Tom Reilly and the people at Cloudera and Hortonworks. On October 2nd they went to bed as two rivals vying for leadership in the big data space. In the morning they woke up as Cloudera 2.0, a $700M firm, with a clear leadership position. “From the edge to AI”…to infinity and beyond!
As an entrepreneur and a CEO of a startup, I think it’s vital to have a mentor. You often come to critical junctures and that’s when it can be invaluable to speak to someone. You can’t talk to your colleagues though because you need to motivate them and keep the organization growing. That’s why having an external mentor is very helpful.
BigQuery is a fast, highly scalable, cost-effective, and fully-managed enterprise data warehouse for analytics at any scale. As BigQuery has grown in popularity, one question that often arises is how to copy tables across locations in an efficient and scalable manner. BigQuery has some limitations for data management, one being that the destination dataset must reside in the same location as the source dataset that contains the table being copied.
Continuing my series of blogs examining the products, services and client benefits borne out of Yellowfin’s partnerships, I’m pleased to introduce our new fintech partner, ChartIQ. ChartIQ provides HTML5 components and the Finsemble integration platform to banks, brokerages, trading platforms, and financial portals worldwide.
I’m seated here on the balcony of our rented apartment in Chamonix looking up at Mont Blanc. It’s an incredible view but its beauty is deceiving and as anyone who has climbed knows, all too often the summit is not the one you can see. In the case of Mont Blanc, the perspective is such that the Dôme du Goûter seen in the foreground often looks bigger. I can assure you it’s not.
When working with tables in data warehouse environments, it is fairly common to come across a situation in which you need to generate surrogate keys. A surrogate key is a system-generated identifier that uniquely identifies a record within a table. Why do we need to use surrogate keys? Quite simply: contrary to natural keys, they persist over time (i.e. they are not tied to any business meaning) and they allow for unlimited values.
Today, Docker or Kubernetes are obvious choices. But, back in 2015, these technologies were just emerging and hoping for massive adoption. How do tech companies make the right open source technology choices early?