Keeping latencies low for highly concurrent, intensive data ingestion
ScyllaDB’s “sweet spot” is workloads over 50K operations per second that require predictably low (e.g., single-digit millisecond) latency. And its unique architecture makes it particularly valuable for the real-time write-heavy workloads such as those commonly found in IoT, logging systems, real-time analytics, and order processing.
Join ScyllaDB technical director Felipe Cardeneti Mendes and principal field engineer, Lubos Kosco to learn about:
- Common challenges that arise with real-time write-heavy workloads
- The tradeoffs teams face and tips for negotiating them
- ScyllaDB architectural elements that support real-time write-heavy workloads
- How your peers are using ScyllaDB with similar workloads
This document discusses big data analytics tools and technologies. It begins with an overview of big data challenges and available tools. It then discusses Packetloop, a company that provides big data security analytics using tools like Amazon EMR, Cassandra, and PostgreSQL on AWS. Next, it discusses how EMR and Redshift from AWS can be used as big data tools for tasks like batch processing, data warehousing, and live analytics. It concludes by discussing how Intel technologies can help power big data platforms by providing optimized processors, networking, and storage to enable analytics at scale.
Amazon RDS for MySQL – Diagnostics, Security, and Data Migration (DAT302) | A...Amazon Web Services
Learn how to monitor your database performance closely and troubleshoot database issues quickly using a variety of features provided by Amazon RDS and MySQL including database events, logs, and engine-specific features. You also learn about the security best practices to use with Amazon RDS for MySQL. In addition, you learn about how to effectively move data between Amazon RDS and on-premises instances. Lastly, you learn the latest about MySQL 5.6 and how you can take advantage of its newest features with Amazon RDS.
Calculating dynamic pricing, estimated travel times or detecting fraud in real time. These are all the cases where realtime analytics create the differentiation between experiences. Redis comes with built in types to enable realtime processing of complex analytics with data types like sorted sets, hyperloglog, bloom and cuckoo filters and more.
The document discusses emerging trends in software and services including:
1) Software as a Service and cloud computing which allows software to be delivered and consumed "as a service" with service level agreements.
2) The growth of massive data centers which are becoming large physical assets requiring significant capital expenditures.
3) The rise of "Dev-signers" or designer-developers who are combining development and design skills.
4) The integration of software and services will be key as local software interacts with internet services to provide combined capabilities.
Hadoop and the Relational Database: The Best of Both WorldsInside Analysis
This document summarizes a presentation about the Splice Machine database product. Splice Machine is described as a SQL-on-Hadoop database that is ACID-compliant and can handle both OLTP and OLAP workloads. It provides typical relational database functionality like transactions and SQL on top of Apache Hadoop. Customers reportedly see a 10x improvement in price/performance compared to traditional databases. The presentation provides details on Splice Machine's architecture, performance benchmarks, customer use cases, and support for analytics and business intelligence tools.
Amazon Redshift is a fast, fully managed data warehousing service that allows customers to analyze petabytes of structured data, at one-tenth the cost of traditional data warehousing solutions. It provides massively parallel processing across multiple nodes, columnar data storage for efficient queries, and automatic backups and recovery. Customers have seen up to 100x performance improvements over legacy systems when using Redshift for applications like log and clickstream analytics, business intelligence reporting, and real-time analytics.
Traditional data warehouses become expensive and slow down as the volume of your data grows. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it easy to analyze all of your data using existing business intelligence tools for 1/10th the traditional cost. This session will provide an introduction to Amazon Redshift and cover the essentials you need to deploy your data warehouse in the cloud so that you can achieve faster analytics and save costs.
This document discusses Bodo Inc.'s product that aims to simplify and accelerate data science workflows. It highlights common problems in data science like complex and slow analytics, segregated development and production environments, and unused data. Bodo provides a unified development and production environment where the same code can run at any scale with automatic parallelization. It integrates an analytics engine and HPC architecture to optimize Python code for performance. Bodo is presented as offering more productive, accurate and cost-effective data science compared to traditional approaches.
AWS Summit 2013 | India - Petabyte Scale Data Warehousing at Low Cost, Abhish...Amazon Web Services
Amazon Redshift is a fast and powerful, fully managed, petabyte-scale data warehouse service in the cloud. In this session we'll give an introduction to the service and its pricing before diving into how it delivers fast query performance on data sets ranging from hundreds of gigabytes to a petabyte or more.
With AWS you can choose the right database technology and software for the job. Given the myriad of choices, from relational databases to non-relational stores, this session provides details and examples of some of the choices available to you. This session also provides details about real-world deployments from customers using Amazon RDS, Amazon ElastiCache, Amazon DynamoDB, and Amazon Redshift.
MongoDB is a document database that provides a more flexible schema than relational databases. It allows embedding related data and easier updates than relational databases with object-relational mapping. MongoDB scales horizontally through sharding and provides high availability through replica sets. It supports different consistency models including eventual and strong consistency through write concerns and read preferences.
Real World Use Cases and Success Stories for In-Memory Data Grids (TIBCO Acti...Kai Wähner
Kai Wähner is a technical lead who discusses in-memory computing and real-world use cases. In-memory computing uses memory for data storage and processing to enable acting in real-time. It offers benefits like eventing, fault tolerance, and high performance beyond traditional caching. Examples where in-memory computing has been applied include personalized customer experiences, routing messages, handling spikes in data, and storing stateful enterprise application data.
Using real time big data analytics for competitive advantageAmazon Web Services
Many organisations find it challenging to successfully perform real-time data analytics using their own on premise IT infrastructure. Building a system that can adapt and scale rapidly to handle dramatic increases in transaction loads can potentially be quite a costly and time consuming exercise.
Most of the time, infrastructure is under-utilised and it’s near impossible for organisations to forecast the amount of computing power they will need in the future to serve their customers and suppliers.
To overcome these challenges, organisations can instead utilise the cloud to support their real-time data analytics activities. Scalable, agile and secure, cloud-based infrastructure enables organisations to quickly spin up infrastructure to support their data analytics projects exactly when it is needed. Importantly, they can ‘switch off’ infrastructure when it is not.
BluePi Consulting and Amazon Web Services (AWS) are giving you the opportunity to discover how organisations are using real time data analytics to gain new insights from their information to improve the customer experience and drive competitive advantage.
The document discusses balancing performance, capacity, and cost for cloud data storage. It notes that most data follows a pattern of occasional reads after initial writes, but some data is frequently read and written. Effective cloud storage needs high capacity and high performance at a lower cost than on-premises storage. While cloud storage was initially just about capacity, it now requires performance for active uses like file synchronization and big data. Performance is needed for cloud computing to be faster and cheaper than alternatives. The document outlines strategies for increasing storage performance like intelligent data placement algorithms and tiering for performance rather than just capacity. This enables cloud providers to reduce costs and increase revenue.
The document discusses balancing performance, capacity, and cost for cloud data storage. It notes that most data follows a pattern of occasional reads after initial writes, but some data is frequently read and written. Effective cloud storage needs high capacity and high performance at a lower cost than on-premises storage. While cloud storage was initially just about capacity, it now requires performance for active uses like file synchronization and big data. Performance is needed for cloud computing to be faster and cheaper than alternatives. The document outlines strategies for increasing storage performance like intelligent data placement algorithms and tiering for performance rather than just capacity. This enables cloud providers to reduce costs and increase revenue.
The document discusses balancing performance, capacity, and cost for cloud data storage. It notes that most data follows a pattern of occasional reads after initial writes, but some data is frequently read and written. Effective cloud storage needs high capacity and high performance at a lower cost than on-premises storage. While cloud storage was initially just about capacity, it now requires performance for active uses like file synchronization and big data. Performance is needed for cloud computing to be faster and cheaper than alternatives. The document outlines strategies for improving performance like intelligent data placement algorithms, tiering for performance rather than just capacity, and using all of a system's performance rather than sacrificing it for capacity.
AWS Cloud Kata | Manila - Getting to Profitability on AWSAmazon Web Services
The document discusses how Lenddo, a financial technology company, has used AWS to scale its operations in a cost-effective manner. It provides details on:
1) How Lenddo started in 2011 in the Philippines and has since expanded to other countries, processing over 50k loan applications for 400k members.
2) How Lenddo's usage of AWS grew significantly from 2011 to 2013 as the company expanded.
3) The various AWS services Lenddo utilizes, including EC2, S3, DynamoDB, RDS, and others, to build its infrastructure in a flexible and scalable way.
4) How using AWS has helped Lenddo focus on coding and
The document discusses new rules and strategies for retailers in an evolving customer relationship landscape. It notes there are now 56 touchpoints between a customer's moment of inspiration and transaction. It then discusses components of digital transformation like customer experience management, cross-channel order orchestration, and building a single customer view. The document outlines how retailers can create customer connections and profiles by leveraging enterprise data. It also discusses the need for customer engagement in stores through technologies like self-scanning and mobile payments. Finally, it discusses how front-end store technologies can empower associates and optimize processes.
This document discusses how to reduce spending on AWS through various techniques:
1. Paying for cloud resources only when they are used through the pay-as-you-go model avoids upfront costs and allows turning off unused capacity.
2. Using reserved instances when capacity needs are predictable provides significant discounts compared to on-demand pricing.
3. Architecting applications in a "cost aware" manner, such as leveraging caching, auto-scaling, managed services, and right-sizing instances can optimize costs.
4. Taking advantage of AWS's economies of scale through consolidated billing and free services helps lower overall spend. Planning workload usage of spot instances can achieve up to 85% savings.
Similar to Using ScyllaDB for Real-Time Write-Heavy Workloads (20)
Unconventional Methods to Identify Bottlenecks in Low-Latency and High-Throug...ScyllaDB
In this presentation, we explore how standard profiling and monitoring methods may fall short in identifying bottlenecks in low-latency data ingestion workflows. Instead, we showcase the power of simple yet clever methods that can uncover hidden performance limitations.
Attendees will discover unconventional techniques, including clever logging, targeted instrumentation, and specialized metrics, to pinpoint bottlenecks accurately. Real-world use cases will be presented to demonstrate the effectiveness of these methods. By the end of the session, attendees will be equipped with alternative approaches to identify bottlenecks and optimize their low-latency data ingestion workflows for high throughput.
Mitigating the Impact of State Management in Cloud Stream Processing SystemsScyllaDB
Stream processing is a crucial component of modern data infrastructure, but constructing an efficient and scalable stream processing system can be challenging. Decoupling compute and storage architecture has emerged as an effective solution to these challenges, but it can introduce high latency issues, especially when dealing with complex continuous queries that necessitate managing extra-large internal states.
In this talk, we focus on addressing the high latency issues associated with S3 storage in stream processing systems that employ a decoupled compute and storage architecture. We delve into the root causes of latency in this context and explore various techniques to minimize the impact of S3 latency on stream processing performance. Our proposed approach is to implement a tiered storage mechanism that leverages a blend of high-performance and low-cost storage tiers to reduce data movement between the compute and storage layers while maintaining efficient processing.
Throughout the talk, we will present experimental results that demonstrate the effectiveness of our approach in mitigating the impact of S3 latency on stream processing. By the end of the talk, attendees will have gained insights into how to optimize their stream processing systems for reduced latency and improved cost-efficiency.
Measuring the Impact of Network Latency at TwitterScyllaDB
Widya Salim and Victor Ma will outline the causal impact analysis, framework, and key learnings used to quantify the impact of reducing Twitter's network latency.
Architecting a High-Performance (Open Source) Distributed Message Queuing Sys...ScyllaDB
BlazingMQ is a new open source* distributed message queuing system developed at and published by Bloomberg. It provides highly-performant queues to applications for asynchronous, efficient, and reliable communication. This system has been used at scale at Bloomberg for eight years, where it moves terabytes of data and billions of messages across tens of thousands of queues in production every day.
BlazingMQ provides highly-available, fault-tolerant queues courtesy of replication based on the Raft consensus algorithm. In addition, it provides a rich set of enterprise message routing strategies, enabling users to implement a variety of scenarios for message processing.
Written in C++ from the ground up, BlazingMQ has been architected with low latency as one of its core requirements. This has resulted in some unique design and implementation choices at all levels of the system, such as its lock-free threading model, custom memory allocators, compact wire protocol, multi-hop network topology, and more.
This talk will provide an overview of BlazingMQ. We will then delve into the system’s core design principles, architecture, and implementation details in order to explore the crucial role they play in its performance and reliability.
*BlazingMQ will be released as open source between now and P99 (exact timing is still TBD)
Noise Canceling RUM by Tim Vereecke, AkamaiScyllaDB
Noisy Real User Monitoring (RUM) data can ruin your P99!
We introduce a fresh concept called ""Human Visible Navigations"" (HVN) to tackle this risk; we focus on the experiences you actually care about when talking about the speed of our sites:
- Human: We exclude noise coming from bots and synthetic measurements.
- Visible: We remove any partial or fully hidden experiences. These tend to be very slow but users don’t see this slowness.
- Navigations: We ignore lightning fast back-forward navigations which usually have few optimisation opportunities.
Adopting Human Visible Navigations provides you with these key benefits:
- Fewer changes staying below the radar
- Fewer data fluctuations
- Fewer blindspots when finding bottlenecks
- Better correlation with business metrics
This is supported by plenty of real world examples coming from the world's largest scale modeling site (6M Monthly visits) in combination with aggregated data from the brand new rumarchive.com (open source)
After attending this session; your P99 and other percentiles will become less noisy and easier to tune!
Always-on Profiling of All Linux Threads, On-CPU and Off-CPU, with eBPF & Con...ScyllaDB
In this session, Tanel introduces a new open source eBPF tool for efficiently sampling both on-CPU events and off-CPU events for every thread (task) in the OS. Linux standard performance tools (like perf) allow you to easily profile on-CPU threads doing work, but if we want to include the off-CPU timing and reasons for the full picture, things get complicated. Combining eBPF task state arrays with periodic sampling for profiling allows us to get both a system-level overview of where threads spend their time, even when blocked and sleeping, and allow us to drill down into individual thread level, to understand why.
Performance Budgets for the Real World by Tammy EvertsScyllaDB
Performance budgets have been around for more than ten years. Over those years, we’ve learned a lot about what works, what doesn’t, and what we need to improve. In this session, Tammy revisits old assumptions about performance budgets and offers some new best practices. Topics include:
• Understanding performance budgets vs. performance goals
• Aligning budgets with user experience
• Pros and cons of Core Web Vitals
• How to stay on top of your budgets to fight regressions
Using Libtracecmd to Analyze Your Latency and Performance TroublesScyllaDB
Trying to figure out why your application is responding late can be difficult, especially if it is because of interference from the operating system. This talk will briefly go over how to write a C program that can analyze what in the Linux system is interfering with your application. It will use trace-cmd to enable kernel trace events as well as tracing lock functions, and it will then go over a quick tutorial on how to use libtracecmd to read the created trace.dat file to uncover what is the cause of interference to you application.
Reducing P99 Latencies with Generational ZGCScyllaDB
With the low-latency garbage collector ZGC, GC pause times are no longer a big problem in Java. With sub-millisecond pause times there are instead other things in the GC and JVM that can cause application threads to experience unexpected latencies. This talk will dig into a specific use where the GC pauses are no longer the cause of unexpected latencies and look at how adding generations to ZGC help lower the p99 application latencies.
5 Hours to 7.7 Seconds: How Database Tricks Sped up Rust Linting Over 2000XScyllaDB
Linters are a type of database! They are a collection of lint rules — queries that look for rule violations to report — plus a way to execute those queries over a source code dataset.
This is a case study about using database ideas to build a linter that looks for breaking changes in Rust library APIs. Maintainability and performance are key: new Rust releases tend to have mutually-incompatible ways of representing API information, and we cannot afford to reimplement and optimize dozens of rules for each Rust version separately. Fortunately, databases don't require rewriting queries when the underlying storage format or query plan changes! This allows us to ship massive optimizations and support multiple Rust versions without making any changes to the queries that describe lint rules.
Ship now, optimize later"" can be a sustainable development practice after all — join us to see how!
How Netflix Builds High Performance Applications at Global ScaleScyllaDB
We all want to build applications that are blazingly fast. We also want to scale them to users all over the world. Can the two happen together? Can users in the slowest of environments also get a fast experience? Learn how we do this at Netflix: how we understand every user's needs and preferences and build high performance applications that work for every user, every time.
Conquering Load Balancing: Experiences from ScyllaDB DriversScyllaDB
Load balancing seems simple on the surface, with algorithms like round-robin, but the real world loves throwing curveballs. Join me in this session as we delve into the intricacies of load balancing within ScyllaDB Drivers. Discover firsthand experiences from our journey in driver development, where we employed the Power of Two Choices algorithm, optimized the implementation of load balancing in Rust Driver, mitigated cloud costs through zone-aware load balancing and combated the issue of overloading a particular core of ScyllaDB. Be prepared to delve into the practical and theoretical aspects of load balancing, gaining valuable insights along the way.
Interaction Latency: Square's User-Centric Mobile Performance MetricScyllaDB
Mobile performance metrics often take inspiration from the backend world and measure resource usage (CPU usage, memory usage, etc) and workload durations (how long a piece of code takes to run).
However, mobile apps are used by humans and the app performance directly impacts their experience, so we should primarily track user-centric mobile performance metrics. Following the lead of tech giants, the mobile industry at large is now adopting the tracking of app launch time and smoothness (jank during motion).
At Square, our customers spend most of their time in the app long after it's launched, and they don't scroll much, so app launch time and smoothness aren't critical metrics. What should we track instead?
This talk will introduce you to Interaction Latency, a user-centric mobile performance metric inspired from the Web Vital metric Interaction to Next Paint"" (web.dev/inp). We'll go over why apps need to track this, how to properly implement its tracking (it's tricky!), how to aggregate this metric and what thresholds you should target.
How to Avoid Learning the Linux-Kernel Memory ModelScyllaDB
The Linux-kernel memory model (LKMM) is a powerful tool for developing highly concurrent Linux-kernel code, but it also has a steep learning curve. Wouldn't it be great to get most of LKMM's benefits without the learning curve?
This talk will describe how to do exactly that by using the standard Linux-kernel APIs (locking, reference counting, RCU) along with a simple rules of thumb, thus gaining most of LKMM's power with less learning. And the full LKMM is always there when you need it!
99.99% of Your Traces are Trash by Paige CruzScyllaDB
Distributed tracing is still finding its footing in many organizations today, one challenge to overcome is the data volume - keeping 100% of your traces is expensive and unnecessary. Enter sampling - head vs tail how do you decide? Let’s look at the design of Sifter and get familiar with why tail-based sampling is the way to enact a cost-effective tracing solution while actually increasing the system’s observability.
Square's Lessons Learned from Implementing a Key-Value Store with RaftScyllaDB
To put it simply, Raft is used to make a use case (e.g., key-value store, indexing system) more fault tolerant to increase availability using replication (despite server and network failures). Raft has been gaining ground due to its simplicity without sacrificing consistency and performance.
Although we'll cover Raft's building blocks, this is not about the Raft algorithm; it is more about the micro-lessons one can learn from building fault-tolerant, strongly consistent distributed systems using Raft. Things like majority agreement rule (quorum), write-ahead log, split votes & randomness to reduce contention, heartbeats, split-brain syndrome, snapshots & logs replay, client requests dedupe & idempotency, consistency guarantees (linearizability), leases & stale reads, batching & streaming, parallelizing persisting & broadcasting, version control, and more!
And believe it or not, you might be using some of these techniques without even realizing it!
This is inspired by Raft paper (raft.github.io), publications & courses on Raft, and an attempt to implement a key-value store using Raft as a side project.
A Deep Dive Into Concurrent React by Matheus AlbuquerqueScyllaDB
Writing fluid user interfaces becomes more and more challenging as the application complexity increases. In this talk, we’ll explore how proper scheduling improves your app’s experience by diving into some of the concurrent React features, understanding their rationales, and how they work under the hood.
The Latency Stack: Discovering Surprising Sources of LatencyScyllaDB
Usually, when an API call is slow, developers blame ourselves and our code. We held a lock too long, or used a blocking operation, or built an inefficient query. But often, the simple picture of latency as “the time a server takes to process a message” hides a great deal of end-to-end complexity. Debugging tail latencies requires unpacking the abstractions that we normally ignore: virtualization, hidden queues, and network behavior.
In this talk, I’ll describe how developers can diagnose more sources of delay and failure by building a more realistic and broad understanding of networked services. I’ll give some real-world cases when high end-to-end latency or elevated failure rates occurred due to factors we ordinarily might not even measure. Some examples include TCP SYN retransmission; virtualization on the client; and surprising behavior from AWS load balancers. Unfortunately, many measurement techniques don’t cover anything but the portion most directly under developer control. But developers can do better by comparing multiple measurements, applying Little’s law, investing in eBPF probes, and paying attention to the network layer.
Understanding API performance to find and fix issues faster ultimately means understanding the entire stack: the client, your code, and the underlying infrastructure.
Ensuring Secure and Permission-Aware RAG DeploymentsZilliz
In this talk, we will explore the critical aspects of securing Retrieval-Augmented Generation (RAG) deployments. The focus will be on implementing robust secured data retrieval mechanisms and establishing permission-aware RAG frameworks. Attendees will learn how to ensure that access control is rigorously maintained within the model when ingesting documents, ensuring that only authorized personnel can retrieve data. We will also discuss strategies to mitigate risks of data leakage, unauthorized access, and insider threats in RAG deployments. By the end of this session, participants will have a clearer understanding of the best practices and tools necessary to secure their RAG deployments effectively.
IVE 2024 Short Course - Lecture 2 - Fundamentals of PerceptionMark Billinghurst
Lecture 2 from the IVE 2024 Short Course on the Psychology of XR. This lecture covers some of the Fundamentals of Percetion and Psychology that relate to XR.
The lecture was given by Mark Billinghurst on July 15th 2024 at the University of South Australia.
Welcome to our third live UiPath Community Day Amsterdam! Come join us for a half-day of networking and UiPath Platform deep-dives, for devs and non-devs alike, in the middle of summer ☀.
📕 Agenda:
12:30 Welcome Coffee/Light Lunch ☕
13:00 Event opening speech
Ebert Knol, Managing Partner, Tacstone Technology
Jonathan Smith, UiPath MVP, RPA Lead, Ciphix
Cristina Vidu, Senior Marketing Manager, UiPath Community EMEA
Dion Mes, Principal Sales Engineer, UiPath
13:15 ASML: RPA as Tactical Automation
Tactical robotic process automation for solving short-term challenges, while establishing standard and re-usable interfaces that fit IT's long-term goals and objectives.
Yannic Suurmeijer, System Architect, ASML
13:30 PostNL: an insight into RPA at PostNL
Showcasing the solutions our automations have provided, the challenges we’ve faced, and the best practices we’ve developed to support our logistics operations.
Leonard Renne, RPA Developer, PostNL
13:45 Break (30')
14:15 Breakout Sessions: Round 1
Modern Document Understanding in the cloud platform: AI-driven UiPath Document Understanding
Mike Bos, Senior Automation Developer, Tacstone Technology
Process Orchestration: scale up and have your Robots work in harmony
Jon Smith, UiPath MVP, RPA Lead, Ciphix
UiPath Integration Service: connect applications, leverage prebuilt connectors, and set up customer connectors
Johans Brink, CTO, MvR digital workforce
15:00 Breakout Sessions: Round 2
Automation, and GenAI: practical use cases for value generation
Thomas Janssen, UiPath MVP, Senior Automation Developer, Automation Heroes
Human in the Loop/Action Center
Dion Mes, Principal Sales Engineer @UiPath
Improving development with coded workflows
Idris Janszen, Technical Consultant, Ilionx
15:45 End remarks
16:00 Community fun games, sharing knowledge, drinks, and bites 🍻
Generative AI technology is a fascinating field that focuses on creating comp...Nohoax Kanont
Generative AI technology is a fascinating field that focuses on creating computer models capable of generating new, original content. It leverages the power of large language models, neural networks, and machine learning to produce content that can mimic human creativity. This technology has seen a surge in innovation and adoption since the introduction of ChatGPT in 2022, leading to significant productivity benefits across various industries. With its ability to generate text, images, video, and audio, generative AI is transforming how we interact with technology and the types of tasks that can be automated.
Flame emission spectroscopy is an instrument used to determine concentration of metal ions in sample. Flame provide energy for excitation atoms introduced into flame. It involve components like sample delivery system, burner, sample, mirror, slits, monochromator, filter, detector (photomultiplier tube and photo tube detector). There are many interference involved during analysis of sample like spectral interference, ionisation interference, chemical interference ect. It can be used for both quantitative and qualitative study, determine lead in petrol, determine alkali and alkaline earth metal, determine fertilizer requirement for soil.
IT market in Israel, economic background, forecasts of 160 categories and the infrastructure and software products in those categories, professional services also. 710 vendors are ranked in 160 categories.
Jacquard Fabric Explained: Origins, Characteristics, and Usesldtexsolbl
In this presentation, we’ll dive into the fascinating world of Jacquard fabric. We start by exploring what makes Jacquard fabric so special. It’s known for its beautiful, complex patterns that are woven into the fabric thanks to a clever machine called the Jacquard loom, invented by Joseph Marie Jacquard back in 1804. This loom uses either punched cards or modern digital controls to handle each thread separately, allowing for intricate designs that were once impossible to create by hand.
Next, we’ll look at the unique characteristics of Jacquard fabric and the different types you might encounter. From the luxurious brocade, often used in fancy clothing and home décor, to the elegant damask with its reversible patterns, and the artistic tapestry, each type of Jacquard fabric has its own special qualities. We’ll show you how these fabrics are used in everyday items like curtains, cushions, and even artworks, making them both functional and stylish.
Moving on, we’ll discuss how technology has changed Jacquard fabric production. Here, LD Texsol takes center stage. As a leading manufacturer and exporter of electronic Jacquard looms, LD Texsol is helping to modernize the weaving process. Their advanced technology makes it easier to create even more precise and complex patterns, and also helps make the production process more efficient and environmentally friendly.
Finally, we’ll wrap up by summarizing the key points and highlighting the exciting future of Jacquard fabric. Thanks to innovations from companies like LD Texsol, Jacquard fabric continues to evolve and impress, blending traditional techniques with cutting-edge technology. We hope this presentation gives you a clear picture of how Jacquard fabric has developed and where it’s headed in the future.
Increase Quality with User Access Policies - July 2024Peter Caitens
⭐️ Increase Quality with User Access Policies ⭐️, presented by Peter Caitens and Adam Best of Salesforce. View the slides from this session to hear all about “User Access Policies” and how they can help you onboard users faster with greater quality.
Leading Bigcommerce Development Services for Online RetailersSynapseIndia
As a leading provider of Bigcommerce development services, we specialize in creating powerful, user-friendly e-commerce solutions. Our services help online retailers increase sales and improve customer satisfaction.
Getting Ready for Copilot for Microsoft 365 with Governance Features in Share...Juan Carlos Gonzalez
Session delivered at the Microsoft 365 Chicago Community Days where I introduce how governance controls within SharePoint Premium are a key asset in a succesfull rollout of Copilot for Microsoft 365. The session was mostly a hands on session with multiple demos as you can see in the session recording available in YouTube: https://www.youtube.com/watch?v=MavcP6k5nU8&t=199s. For more information about Governance controls available in SharePoint Premium visit official documentation available at Microsoft Learn: https://learn.microsoft.com/en-us/sharepoint/advanced-management
Connecting Attitudes and Social Influences with Designs for Usable Security a...Cori Faklaris
Many system designs for cybersecurity and privacy have failed to account for individual and social circumstances, leading people to use workarounds such as password reuse or account sharing that can lead to vulnerabilities. To address the problem, researchers are building new understandings of how individuals’ attitudes and behaviors are influenced by the people around them and by their relationship needs, so that designers can take these into account. In this talk, I will first share my research to connect people’s security attitudes and social influences with their security and privacy behaviors. As part of this, I will present the Security and Privacy Acceptance Framework (SPAF), which identifies Awareness, Motivation, and Ability as necessary for strengthening people’s acceptance of security and privacy practices. I then will present results from my project to trace where social influences can help overcome obstacles to adoption such as negative attitudes or inability to troubleshoot a password manager. I will conclude by discussing my current work to apply these insights to mitigating phishing in SMS text messages (“smishing”).
Multimodal Embeddings (continued) - South Bay Meetup SlidesZilliz
Frank Liu will walk through the history of embeddings and how we got to the cool embedding models used today. He'll end with a demo on how multimodal RAG is used.
The Challenge of Interpretability in Generative AI Models.pdfSara Kroft
Navigating the intricacies of generative AI models reveals a pressing challenge: interpretability. Our blog delves into the complexities of understanding how these advanced models make decisions, shedding light on the mechanisms behind their outputs. Explore the latest research, practical implications, and ethical considerations, as we unravel the opaque processes that drive generative AI. Join us in this insightful journey to demystify the black box of artificial intelligence.
Dive into the complexities of generative AI with our blog on interpretability. Find out why making AI models understandable is key to trust and ethical use and discover current efforts to tackle this big challenge.
Project Delivery Methodology on a page with activities, deliverablesCLIVE MINCHIN
I've not found a 1 pager like this anywhere so I created it based on my experiences. This 1 pager details a waterfall style project methodology with defined phases, activities, deliverables, assumptions. There's nothing in here that conflicts with commonsense.
Project Delivery Methodology on a page with activities, deliverables
Using ScyllaDB for Real-Time Write-Heavy Workloads
1. Using ScyllaDB for
Real-Time Write-Heavy
Workloads
Felipe Cardeneti Mendes, Technical Director, ScyllaDB
Lubos Kosco, Principal Field Engineer, ScyllaDB
2. + For data-intensive applications that require high
throughput and predictable low latencies
+ Close-to-the-metal design takes full advantage of
modern infrastructure
+ >5x higher throughput
+ >20x lower latency
+ >75% TCO savings
+ Compatible with Apache Cassandra and Amazon
DynamoDB
+ DBaaS/Cloud, Enterprise and Open Source
solutions
The Database for Gamechangers
2
“ScyllaDB stands apart...It’s the rare product
that exceeds my expectations.”
– Martin Heller, InfoWorld contributing editor and reviewer
“For 99.9% of applications, ScyllaDB delivers all the
power a customer will ever need, on workloads that other
databases can’t touch – and at a fraction of the cost of
an in-memory solution.”
– Adrian Bridgewater, Forbes senior contributor
3. 3
+400 Gamechangers Leverage ScyllaDB
Seamless experiences
across content + devices
Digital experiences at
massive scale
Corporate fleet
management
Real-time analytics 2,000,000 SKU -commerce
management
Video recommendation
management
Threat intelligence service
using JanusGraph
Real time fraud detection
across 6M transactions/day
Uber scale, mission critical
chat & messaging app
Network security threat
detection
Power ~50M X1 DVRs with
billions of reqs/day
Precision healthcare via
Edison AI
Inventory hub for retail
operations
Property listings and
updates
Unified ML feature store
across the business
Cryptocurrency exchange
app
Geography-based
recommendations
Global operations- Avon,
Body Shop + more
Predictable performance for
on sale surges
GPS-based exercise
tracking
Serving dynamic live
streams at scale
Powering India's top
social media platform
Personalized
advertising to players
Distribution of game
assets in Unreal Engine
4. Presenters
Felipe Cardeneti Mendes, Technical Director
+ Puppy Lover
+ Open Source Enthusiast
+ ScyllaDB passionate!
Lubos Kosco, Principal Field Engineer
+ Software Engineer
+ 🎹 lover, 🪂pilot
+ ScyllaDB enthusiast :-)
7. + Commonly referred to as "write-mostly"
+ Workloads requiring high volume of writes under very low response times
+ Challenges involve:
+ Scaling writes – Price per operation, often driven by internal design decisions
+ Locking – Add delays and reduce throughput
+ I/O Bottlenecks – Write amplification & Crash Recovery
+ Conflict Resolution – Resolving conflicts and/or Commit Protocols
+ Database Backpressure – Throttling incoming load
Real-time Write-Heavy?
8. + 📟 Internet of Things
+ Often time-series workloads
+ Small (but frequent!) append-only writes
+ Rate determined by number of ingestion endpoints
+ 📶 Logging and Monitoring
+ Similar to IOT, but…
+ Doesn't have a fixed ingestion rate
+ Not necessarily append-only, and prone to hotspots
+ 🎮 Online Gaming
+ Real-time user interactions (state, actions, messaging)
+ Often spiky
+ Very latency dependant
Commonly Seen Use Cases
+ 🚚 E-commerce & Retail
+ Update-heavy and Batch-y
+ Inventory updates, reviews, order status & placement
+ Shopping carts inherently require a read-before-write
+ 🔔 Ad Tech and Real-time Bidding
+ Bid Processing (Impressions, Auction outcomes)
+ User Interaction (Clicks, Conversions, Fraud Detection)
+ Audience Segmentation
+ 📈 Real-time Stock exchange
+ High-frequency trading (HFT)
+ Stock prices updates
+ Order matching
14. Compression Chunk Size
Selecting Compression Chunk Sizes for ScyllaDB
+ Determines the size of a compression block
+ ScyllaDB block size defaults to: 4kB (SSTable), 1MB (RAID), filesystem sector size 1kB || 4kB
+ Trade-off:
+ Larger chunk sizes – Reduces the bandwidth used to write data
+ Smaller chunk sizes – Reduces the bandwidth needed to read data
Chunk size > Partition Size
Chunk size ~= Partition Size
Use case Recommendation Comments
small single key smaller chunks close to partition size
large single key larger chunks close to partition size
partition scans larger chunks good cache locality
mostly writes larger chunks saves write bandwidth
15. Compaction Strategy
+ The goal of a compaction strategy is low amplification
+ Read amplification – Avoid reading from too many SSTables
+ Write amplification –Avoid re-writing the same data over and over again
+ Space amplification – Avoid expired/deleted/overwritten data sitting on disk for too long
+ If write performance is important for you…
+ ❌ Avoid Leveled Compaction at all costs!
Every byte has to be rewritten up to 10 times per level
4 Levels usually - potentially up to 40x write amplification
20. Batching – Good Pattern
Client App
All to the same partition
Christopher Batey's – Misuse of unlogged batches
21. Views & Global Indexes
+ All writes to a base table are eventually propagated to the view table
+ If the update changes a view’s key column, this mean deleting an old view row and
creating new one:
UPDATE tbl SET x=3
WHERE pk=1
DELETE FROM view
WHERE x=<old value>
INSERT INTO view (x, …)
VALUES (3, <old data>)
read-before-write
22. Local Indexes, CDC & Other Animals
+ All writes are applied as a single mutation, however…
+ It also results in write amplification as it requires yet another write
UPDATE tbl SET x=3
WHERE pk=1
INSERT INTO index (pk, x,
…) VALUES (1, 3)
INSERT INTO tbl_cdc_log (…)
VALUES (…)
24. Zillow – Real-time Property Updates
Consuming records from different data producers: Out of Order Writes
+ Why not SQL? Locking.
+ Can't simply insert everything: Data bloat, higher costs
+ Solution: INSERT … USING TIMESTAMP ?
“No one will even notice that we’re processing the entirety of Zillow’s property and listings data in order
to correct some data issue or change a business rule. The beauty of that is we can process the entirety
of the data at Zillow that we care about in less than a business day and, again, no performance hit to
real-time data.” – Dan Podhola, Principal Software Engineer
Zillow: Optimistic Concurrency with Write-Time Timestamps
25. Unleashing the Power of Data with Scylla
"We were looking for something
really specific.A highly scalable,
and highly performant NoSQL
database.
The answer was simple,
ScyllaDB is a better fit for our
use case."
João Pedro Voltani – Head of Engineering
26. Fanatics – Retail Operations
Use cases:
+ Order capture
+ Product Catalog
+ Shopping Carts
+ Promotions, …
“During a recent peak minute, we saw nearly 280,000 IOPs for a solid minute. With a 3-node ScyllaDB
cluster, we registered zero timeouts. Because of this we had happier customers and application
teams.” – Niraj Kothari, Director of Platforms Engineering
Read the Case Study
27. Poll
How much data do you have under
management of your transactional
database?