Organizations are sitting on a mountain of data and untapped business intelligence, all stored across various internal and external systems. Those that utilize their data and analytics the best and the fastest will deliver more revenue, better customer experience, and stronger employee productivity than their competitors. And the organizations that can answer the tough questions […]
Only a few years ago, IT teams and their security counterparts worked in separate silos. But as organizations face escalating threats and breaches, there is a burning need to converge IT observability and cybersecurity. The single biggest driver of this convergence is the need to share critical data to help security teams improve cyber resilience. […]
In my view, effective leadership involves empowering employees and fostering a culture of innovation amongst teams. This is particularly important in industry sectors that are bound up in imprecise ways to achieve both specific and vaguely specified outcomes. These may include technology and software development, entertainment and media, fashion and design, advertising and marketing, renewable energy, and sustainability.
The concept of autonomous yet thoughtful decision-making is a powerful leadership strategy that helps to drive desired productive outcomes. Many may understand the significance of autonomy and empowerment but not acknowledge the importance in various business settings.
This often means emphasizing the need for a shift away from very traditional command and control models (C&C). C&C is prevalent in more traditional and bureaucratic organizations it has often been associated with industries where standardization, efficiency, and compliance were crucial, such as military, manufacturing, and certain government sectors.
Some of the key characteristics of C&C include centralized decision-making where the decision-making power is concentrated in the hands of those at the top. This approach often leaves little room for input from the lower levels of employees. There’s a chain of command and decisions are typically passed down that chain.
A second common characteristic is the strictness of the hierarchy. Organizationally the structure is typically like a pyramid with clearly delineated lines of authority and control. Each level reports to the one above it, and instructions flow from the top down. There may be an emphasis on discipline and control to ensure that employees adhere to prescribed and somewhat predictable processes in order to meet performance expectations.
C&C is often characterized by rigid adherence to rules, procedures, and protocols. Employees are expected to follow specific guidelines as prescribed without deviation. In line with the pyramid, communication follows formal channels, such as through managers and supervisors and information or insights may be limited as the communication slows up and down the hierarchy. Everyone is assigned specific roles and responsibilities, and tasks are somewhat clearly defined. Employees have little autonomy to make decisions or exercise creativity. The focus is on carrying out assigned tasks as directed.
While this leadership model can be effective in certain circumstances as I previously described, it is often criticized for its limitations in an ambiguous, dynamic, and fluid business environment.
In industries that require adaptability, creativity, and innovation, like the tech sector, the command and control model in my experience, actually hinders employee engagement, limits the flow of ideas, and inhibits organizational agility. I am more in favor of participative and collaborative leadership that empowers employees and fosters a culture of innovation and a genuine desire for ownership and accountability.
Instead, I advocate for a more informal and relaxed, and collaborative leadership approach that encourages creativity and innovation where the leadership team functions as player-coaches and builds genuine consensus and collective agreement on big and small decisions through dialog and negotiation.
Growth through empowerment
If you want growth, then empowering employees goes beyond simple delegation; it requires trusting individuals to make informed decisions and providing them with the necessary resources and autonomy to act. In so doing you foster a sense of ownership and accountability within the workforce which then leads to higher job satisfaction and improved overall productivity.
The core of successful empowerment lies in striking the right balance between autonomy and thoughtful decision-making. You should want autonomous yet well-considered decision-making from your employees. Autonomy allows teams and individuals to leverage their expertise and creativity to address complex challenges effectively. However, it must be complemented with considered decision-making, where employees gather information, seek advice, and analyze potential outcomes before acting. Remember you’re paying these people for their expertise and to perform a particular job. If you’re only interested in barking instructions at them then you may as well just hire unskilled people with no particular specialisms or experience.
Tailoring empowerment models to your business settings is important since the universal benefits of empowerment and autonomy don’t necessarily manifest in all cultures or work settings. The application should therefore be tailored to suit the specific business contexts you have. There are a few different implementation models to consider. Task-based, Team-based, and Individualized
Task-based empowerment is typical for industries with routine tasks. Task-based empowerment can streamline processes and enhance productivity. By granting employees authority over specific responsibilities, business leaders enable them to make decisions related to their assigned tasks, boosting efficiency. Without disrupting efficiency and effectiveness, for example, employees can rotate and resequence their tasking according to their preferences and observed optimality.
Team-based empowerment is most appropriate in dynamic environments which ultimately benefit from improved collaboration and collective decision-making and where these activities take center stage. By allowing teams to pool diverse perspectives and expertise, leaders have the potential to tap into the opportunities afforded by collective innovation.
In roles requiring specialized skills, individual-based empowerment can be highly effective. Leaders empower subject matter experts to make decisions in their areas according to proficiency, fostering innovation and excellence in technology-driven fields. This is individual-based empowerment
C&C with its centralized decision-making and strict protocols works somewhat well in highly regulated industries, this approach stifles creativity though and limits adaptability in technology development. Employees may feel restricted, resulting in decreased motivation, innovation, and engagement.
Conversely, the informal and relaxed leadership style promotes open communication, trust, and collaboration. By empowering employees to make autonomous decisions, leadership fosters an essential culture of innovation and agility. This approach is particularly effective in software development and technology-driven operations, where creativity thrives in a flexible environment.
Getting the best out of teams and individuals
Getting the best out of the two quite different approaches still requires you to set clear objectives with agreed measurable outcomes. Empowering employees requires providing clear objectives and expectations. Well-defined goals using the SMART framework (Specific, Measurable, Achievable, Relevant, Time-bound) help individuals self-manage their work efficiently.
Another facet is effective time management; autonomy allows individuals to manage their own time. To be effective though, there needs to be discipline. Discipline is essential for ensuring its effective use. Encouraging employees to prioritize tasks, set deadlines, and avoid distractions maintains their productivity.
Autonomous employees must also be accountable for their work. Encouraging ownership and transparent progress reporting foster a sense of responsibility.
Self-Management in Project and Program Management
In technology development, adopting an agile methodology enhances self-management. Empowered teams self-organize, collaborate, and make quick decisions to adapt to changing requirements effectively.
Leaders can further empower teams by providing autonomy in decision-making. Open communication and input from team members drive a self-managed and collaborative environment.
Project management itself involves ongoing learning and improvement allowing employees to reflect on progress and take initiative. These empowering approaches support positive change and have a greater likelihood of driving success.
As already suggested, empowerment also requires balancing discipline with flexibility. Through research, it has been found that Innovation thrives in more flexible environments. Leaders must therefore be open to diverse methods and ideas, trusting teams to find effective solutions. Open channels of communication facilitate not only bidirectional trust, they also support employee self-management and lead to continuous improvement.
A few words of caution
Sometimes, in our earnest efforts to empower others and provide autonomy, we may inadvertently deceive ourselves into believing that we have relinquished command and control.
Despite statements of intent, the empowerment we claim to have granted is a self-deceiving illusion. We might unknowingly perpetuate a micro-management overreach by behaving in exactly the opposite way to what we suggest we are thinking. This can occur when we continuously bombard teams with questions, second-guess their independent decisions, and challenge their judgment with frequency. Individuals and teams need to occasionally fail to learn. While our intention may be to offer support and ensure success, our actions may inadvertently stifle creativity and autonomy.
True empowerment necessitates trust and allowing individuals the space to take ownership of their work is critical. Constantly questioning decisions sends mixed signals, who is actually making the decisions? Undermining the confidence of team leads and team members impedes their ability to innovate freely. To genuinely empower others, we must genuinely let go of control, offer guidance only when sought, celebrate their success, recognize missteps, and offer encouragement, coaching, and reassurance.
Occasional mistakes are part of the learning process. By fostering a culture of trust and granting autonomy, we can break free from the C&C mindset, unleashing the full potential of our teams to drive creativity and achieve remarkable results.
Nurturing a culture of empowerment is essential for fostering innovation in business leadership. By tailoring empowerment models to the specifics of the business setting and adopting informal leadership styles, leaders can cultivate creativity and adaptability, particularly in software development and technology-driven operations. Encouraging discipline and self-management in tasking and project and program management enables employees to thrive in an environment that values autonomy while maintaining focus and efficiency. Striking the right balance between discipline and flexibility empowers teams to innovate, drive success, and contribute to sustainable growth.
While AI’s ubiquity is becoming increasingly evident through everyday tools like chatbots, smart cameras, and smart content generation, there’s an expansive universe of less recognized but highly potent advancements poised to redefine how data scientists interact with and leverage the burgeoning volume and complexity of datasets. Emerging AI trends such as natural language processing, reinforcement learning, […]
As a returning intern who has spent two memorable summers at Actian, I am excited to share what sets Actian apart and how their internship program is truly exceptional.
One aspect of the program I absolutely adore is the in-person orientation. Unlike many remote companies, Actian offers a unique opportunity for interns to come together at the Center of Excellence office in Round Rock, Texas.
Trust me when I say this is one of the highlights of the summer! Not only do you get to bond with fellow interns and spend quality time with your manager and buddy, but you also get to explore Texas together and engage in exciting activities. A personal favorite of mine was the gameshow-style intern competition at GameOn! ATX. Despite my team losing for the second year in a row, we gave it our all, making it an exciting and fun event.
While some worry that you forfeit feelings of community when working remotely, that is never a problem at Actian. Laughter and jokes are a regular part of our meetings, and it’s normal to see funny GIFs in our team chats. Whether it’s work anniversaries or birthdays, we celebrate the little things together. Last summer, the Actian team went above and beyond by surprising me with an edible arrangement and cupcakes to celebrate my birthday. These gestures truly exemplify the caring and supportive atmosphere within the company.
I must also mention the remarkable Marketing team, not only because I am a marketing intern, but because the people at Actian are genuinely one of a kind. You cannot find a more collaborative, enthusiastic, and supportive group. Despite my short tenure as a summer intern, I even have my own nickname (the Sparkler) within the marketing team! Across all departments, Actian employees are eager to connect with you, help you learn, and foster your professional growth.
Actian truly makes interns feel valued and appreciated, and I couldn’t be more grateful to be here another summer. I will carry the memories, experiences, and relationships I’ve built at Actian long after my internship ends!
Returning intern, Madeline, is pursuing majors in Marketing and International Business at Indiana University Bloomington’s Kelley School of Business. At Actian, she assists the marketing team with various projects but is primarily focused on the social gamification of the employee advocacy platform to leverage employee networks and achieve brand visibility and thought leadership. She cannot say enough about the people and culture at Actian!
Savannah is a recent graduate of Loyola University Chicago where she majored in in Statistics and Public Relations. At Actian, she develops and executes A/B, Multivariate, and continuous optimization tests, takes winning ideas to market and integrates them into the Actian website! Her favorite part of the internship is applying skills she learned in the classroom and learning new skills!
Caitlin studies Information Systems at Fordham University. As a Cloud Operations Analyst intern, she works alongside her project manager to ensure efficiency of the Cloud Ops team. She enjoys collaborating with her buddy/manager on her Capstone project and the weekly intern events!
Ian Loo – DataConnect Engineering Intern
Ian is majoring computer science at California Polytechnic State University, San Luis Obispo. At Actian, he carries out specific tasks for the DataConnect team through investigation, brainstorming, and implementation. His favorite part of his internship is learning new technologies and industry practices!
Alexa Cole – DevOps Intern
Alexa is majoring in computer engineering at the University of Florida. As the Actian DevOps intern, she works to perform migration from Twiki to Confluence. She has loved learning about what DevOps is truly about!
Ethan Avila – Cloud Security Engineer Intern
Ethan is majoring in Technology Management at Texas A&M University. At Actian, Ethan is learning how to build, maintain, and enhance protection and detection on cloud-based infrastructure. He is grateful for how much he has been able to learn in just a short time!
Henry McKinney – Cloud Operations Analyst Intern
Returning intern, Henry, is a recent graduate of Nazareth University where he majored in psychology and triple-minored in math, analytics, and finance. As a member of the Customer Cloud Operations team, Henry utilizes Kubernetes to manage cloud native applications. He has loved learning about Kubernetes this summer!
Sayali Dalvi – DataConnect Engineering Intern
Sayali is majoring in Information Systems at Northeastern University. As Actian’s DataCloud Engineering, she is working to build new features for the Integration of Actian’s flagship product, Avalanche. She looks forward to the weekly intern events and weekly 1:1 meeting with the employee experience team! She really appreciates the personalized guidance and feedback!
Phuong Tran – Employee Experience Intern
Phuong is pursuing a master’s in industrial and organizational Psychology at the University of Central Florida. As an Employee Experience Intern, her primary responsibility entails fostering global corporate volunteering amongst Actian employees by establishing a corporate volunteering program that reflects Actian’s values of corporate social responsibility (CSR). She loves the weekly intern events and enjoys expanding her knowledge of the industry through real-world application!
Gabriela Flores Diaz – Online Media Review and ESG Intern
Gabriela is double majoring in marketing and communication at the Arizona State University W.P. Carey School of Business. In her current role, she is developing a new online media review process and implementing an ESG marketing strategy. She has loved connecting with fellow interns and members of the Actian team, especially during in person orientation! The collaboration with professionals from various functions at Actian has inspired her with valuable insights and diverse perspectives.
Matthew Jackson – Zen Engineering Intern
Returning intern, Matthew Jackson is majoring in Computer Science at the Colorado School of Mines. As the Zen Engineering intern, Matthew works to create a standalone utility application to assist developers with integrating Zen and working out a sample to demonstrate a possible Zen use case. He loves learning more about the industry through hands-on applications and working with incredible people!
Skylar Ng – Employee Experience Intern
Skylar is majoring in Psychology at Michigan State University. As an Employee Experience Intern, he primarily focuses on helping Actian on its path to sustainability through research and policy implementation. He loves connecting with other interns and employees across Actian and gaining insight into their roles and responsibilities.
Aditi is majoring in Economics and Urban and Environmental Policy at Occidental College. As a Sustainability intern, she researches and collaborates with the engineering team to ensure Actian’s software is environmentally friendly and works to solidify Actian’s ESG mission! She has loved getting to and working with her fellow interns!
Sahithi Eslavath – Software Engineering Intern
Sahithi is majoring in Data Science at Greenwich University. As an Actian Software Engineering Intern, she works to establish the connection from DBT to Ingres Avalanche. During her time at Actian, she has loved developing the new adapter
Josh Reyes – Financial Planning & Analysis Intern
Returning intern, Josh, is majoring in Finance at the University of San Francisco. As Actian’s Financial Planning & Analysis intern, he creates dashboards to track finances and company metrics. He loves the Actian intern program and really enjoys working with the finance team, learning about the technology industry, and bonding with his fellow interns!
A comprehensive customer master data management strategy and approach to good customer data governance are essential for any organization that wants to effectively manage and leverage its consumer data.
This is particularly important for sales, marketing, and service departments that interact directly with consumer customers.
A Customer Master Data Management (CMDM) solution helps manage and maintain a single, consistent view of customer data across the organization. This can include capabilities such as hierarchy management, data modeling, and governance.
But, the Chief Data Officer (CDO) has many product choices to consider as part of the customer data management strategy apart from a Customer Master Data Management solution. Here are a few to consider:
Data Quality Tools help ensure that customer data is accurate, complete, and consistent. They can be used to validate, standardize and cleanse data, as well as to identify and resolve any data quality issues.
Customer data often contains errors, duplicates, and inconsistencies. Data cleansing tools help identify and correct these issues, ensuring data accuracy and reliability. Deduplication tools enable the identification and merging of duplicate customer records, eliminating redundancy and improving data quality.
The Pretectum CMDM also provides some lightweight data quality capabilities including deduplication at the time of data loading and during customer data lifecycle management.
Data Integration Tools help integrate customer data from various sources into a single, unified view. This can include data from internal systems such as CRM and ERP systems, as well as external sources such as social media and third-party data providers.
Integration tooling can assist in the key areas of customer data integration; if you need to combine data from different sources into a unified format and structure. Data integration can also aid data extraction, transformation, and loading (ETL), ensuring that it is harmonized and consistent across systems.
The Pretectum CMDM provides you with various integration options that facilitate integration between systems, particularly if you choose to use the Pretectum CMDM as the data hub.
Data enrichment tools allow you to enhance customer profiles by appending additional information from external sources. This can include demographic data, social media data, purchasing behavior data, and more. Data enrichment enables a deeper understanding of customers and supports personalized interactions.
Data Governance Tools help establish and enforce policies and procedures for managing customer data. This can include capabilities such as access control, audit trails, and data lineage.
Data governance frameworks provide guidelines, policies, and processes for managing customer data. They suggest approaches to compliance with data privacy regulations, established data ownership, defined data access controls, and managed data lifecycle. Data governance tools enable the enforcement of data policies, monitoring of data usage, and tracking of data lineage.
Data Stewardship means rethinking your organization structures and often taking the bold step of selecting and assigning data stewards; stakeholders who are responsible for overseeing the quality, integrity, and governance of customer data is an important step towards better customer data management.
Stewards should establish data standards, enforce data policies, and drive data-related initiatives across the organization. Implementing robust data security measures, including encryption, access controls, and monitoring, to protect customer data from unauthorized access, breaches, or theft along the way.
Configuring solutions like the Pretectum CMDM is an essential part of this process.
Data Privacy and Compliance policies protect customer data and ensure compliance with relevant regulations, such as GDPR or CCPA. Establishing policies and procedures for data handling, access control, and consent management can be reinforced through specific tools and the Precetum CMDM can be leveraged to augment or support such policies.
Depending on how you implement the Pretectum CMDM, the sophisticated Role Based Access Controls allow you to control access, provide audit trails, and also support data lineage views of sources and targets.
Analytics and Business Intelligence (BI) Tools: These tools help analyze customer data to gain insights and make informed decisions. This can include capabilities such as reporting, dashboards, and predictive analytics.
Customer analytics tools, in particular, leverage advanced analytics techniques, such as machine learning and predictive modeling, to gain insights into customer behavior, preferences, and patterns. These tools enable segmentation, trend identification, and data-driven decision-making, empowering sales, marketing, and service teams to deliver personalized experiences and targeted campaigns.
Pretectum’s CMDM provides some lightweight reporting capabilities but also supports your ability to access your Customer Master Data repository in controlled ways for using your own reporting tools.
It’s important to note that a successful customer master data management strategy involves not only the right tools and products but also the right processes, people, and culture. A strong focus on data governance, collaboration between departments, and a commitment to continuous improvement are all essential for achieving success.
To effectively manage customer data from these diverse sources, you need to leverage the most appropriate data tools, policies, procedures, frameworks, methods, and products; Pretectum think that a CMDM is one of those data tools.
Increasingly more organizations initiate AWS migration due to the advantages provided by the AWS cloud platform. According to The Business Value of Migrating to AWS survey, 43% of AWS adopters report lower time-to-market for new features, 20% report infrastructure cost savings, and 66% note an increase in administrator productivity. However, AWS migration is a challenging task requiring a […]
In my career, I’ve had first-hand experience as both a user and a chooser of data analytics technology, and have also had the chance to talk with countless customers about their data analytics journey to the cloud. With some reflection, I’ve distilled the learnings down to 6 key questions that every technology and business leader should ask themselves to avoid pitfalls along the way to the cloud so they can achieve its full promise.
1. What is my use case?
Identifying your starting point is the critical first step of any cloud migration. The most successful cloud migrations within our customer base are associated with a specific use case. This focused approach puts boundaries around the migration, articulates the desired output, and enables you to know what success looks like. Once a single use case has been migrated to the cloud, the next one is easier and often relies on data that has already been moved.
2. How will we scale over time?
Once you’ve identified the use case, you’ll need to determine what scaling looks like for your company. The beauty of the cloud is that it’s limitless in its scalability; however, businesses do have limits. Without planning for scale, businesses run the risk of exceeding resources and timelines.
To scale quickly and maximize value, I always recommend customers evaluate use cases based on level of effort and business value: plotting each use case in a 2×2 matrix will help you identify the low effort, high value areas to focus on. By planning ahead for scale, you de-risk the move to the cloud because you understand what lies ahead.
3. What moves, what doesn’t, and what’s the cost of not planning for a hybrid multi-cloud implementation?
We hear from our customers, especially those in Europe, that there is a need to be deliberate and methodical in selecting the data that moves to the cloud. Despite the availability of data masking, encryption, and other protective measures available, concerns about GDPR and privacy are still very real. These factors need to be considered as the cloud migration roadmap is developed.
Multi-cloud architectures create resiliency, address regulatory requirements, and help avoid the risk of vendor lock-in. The benefits of multi-cloud environments were emphasized in a recent meeting with one of our EMEA-based retail customers. They experienced significant lost revenue and reputation damage after an outage of one of the largest global cloud service providers. The severe impact of this singular outage made them rethink a single cloud strategy and move to multi-cloud as part of their recovery plan.
4. How do I control costs?
In our research on customers’ move to the cloud, we found that half of organizations today are demanding better cost transparency, visibility, and planning capabilities. Businesses want a simple interface or console to determine which workloads are running and which need to be stopped – the easier this is to see and control, the better. Beyond visibility in the control console, our customers also use features such as idle stop, idle sleep, auto-scaling, and warehouse scheduling to manage costs. Every company should evaluate product performance and features carefully to drive the best cost model for the business. In fact, we’ve seen our health insurance customers leverage performance to control costs and increase revenue.
5. What skills gaps will I need to plan for, and how will I address them?
Our customers are battling skills gaps in key areas, including cloud, data engineering, and data science. Fifty percent of organizations lack the cloud skills to migrate effectively to the cloud, and 45 percent of organizations struggle with data integration capacity and challenges, according to our research. Instead of upskilling a team, which can often be a slow and painful process, lean on the technology and take advantage of as-a-service offerings. We’ve seen customers that engage in services agreements take advantage of platform co-management arrangements, fully managed platform services, and outsourcing to help offset skills gap challenges.
6. How will I measure success?
Look beyond cost and measure success based on the performance for the business. Ask yourself: is your cloud solution solving the problem you set out to solve? One of our customers, Met Eireann, the meteorological service for Ireland, determined that query speed was a critical KPI to measure. They found after moving to the cloud that performance improved 60-600 times and reduced query result time down to less than a second. Every customer measures success differently, whether it’s operational KPIs, customer experience, or data monetization. But whatever the measure, make sure you define success early and measure it often.
Making the move to the cloud is a journey, not a single step. Following a deliberate path, guided by these key questions, can help you maximize the value of cloud, while minimizing risk and disruption. With the right technology partner and planning, you can pave a smooth road to the cloud for your organization and realize true business value from your data.
Data analysts are an important part of modern business, and this article will not argue otherwise. But it will argue that not every company, department, and team needs a data analyst to become data-driven. Before being hired as senior product manager of a data integration platform, I had little experience working directly with data tools. But I knew […]
GitOps is a way of implementing continuous delivery for cloud native applications. It is based on the idea of using Git as a single source of truth for declarative infrastructure and applications. In GitOps, the desired state of the infrastructure and applications is stored in version control, and an automated process is used to ensure […]
Factors Influencing Decision-Making in Parity Competitive Products
It is interesting when you work in a competitive technology space, exploring how decisions are arrived at, especially in terms of customer technology choices.
As individuals, we face these challenges regularly, perhaps not even really thinking about the process much. We do it, by choosing appliances, mobile phones, cars, houses, etc. Our choices and decisions about products are influenced by a complex interplay of cognitive, emotional, social, and situational factors.
Researchers delve ever deeper into trying to understand these dynamics to help businesses create more effective marketing strategies and to aid policymakers in promoting informed decision-making among consumers but in the end, we don’t seem to really have a magic formula for how we settle on a decision.
In the corporate world, the same challenge for consumers is of selecting the most suitable technology solutions to meet their specific needs and objectives. Whether it’s for software, hardware, or other IT solutions, the decision-making process is often complex and critical for success, both personally, and for the organization.
In cases where competing technologies exhibit similar functionality and capabilities, additional factors become crucial in influencing the final selection. Consider the significance of various factors in the decision-making process, including the character and personality of the customer account manager, the presales consultant, the engagement process, pricing, company size, customer references, and other pertinent aspects, all of which, in my experience, influence the final outcome.
Picking and choosing
Organizationally, defining a need and a budget is often the starting point. Something doesn’t work, something is broken, or something is problematic. The idea might be to automate, it may be to replace or renew. When organizations seek particular technology solutions they will typically define requirements and objectives clearly.
The framing of the problem is often part of the budget-requesting process but sometimes the solution is already conceived and the problem or problem statement is already relegated. Appropriate requirements definition involves understanding the specific problems to be solved, suggesting the desired outcomes, and trying to arrive at the key performance indicators (KPIs) that will be used to measure the success of the chosen solution. If you don’t have a problem statement and you don’t have success measures then you likely don’t have a clear vision.
This may seem incredibly obvious, but if we again revert to thinking like a consumer, if you want to replace your refrigerator, you must be clear as to why you want to replace it. You may have experienced a windfall, but what is the problem with the fridge you have? You may have moved, and have no room for your old fridge or no fridge at all, or your existing fridge may actually be defective or broken. Your problem statement likely isn’t “I have money to burn – I will buy a fridge” it has to be something else. As an organization there has to be a clear vision about what problems are to be solved through the acquisition, this in turn initiates an evaluation and selection process.
Requests
For big corporate purchases, there is the RFI, RFP, RFQ process. Some have all of these, some have just a couple of these and the thresholds vary according to the specific purchasing policies.
An organization may start by issuing a Request for Information (RFI) to potential vendors. This initial step helps gather general information about products and solutions available in the market. The RFI allows the vendor to provide high-level overviews of offerings, giving the organization a broad understanding of what is available. The trigger for an RFI is often off the back of preliminary or cursory research. Using industry analyst insights, using comparison sites, or simply familiarity with particular technology vendors.
After the RFI phase (which might be skipped altogether), the organization may choose to issue a Request for Proposal (RFP) or Request for Quotation (RFQ) to a select group of vendors. These documents contain detailed specifications, requirements, and evaluation criteria.
Vendors respond with comprehensive proposals outlining how their technology solutions align with the organization’s needs. Invited participants may need to sign non-disclosure agreements (NDA) or may choose to ignore the RFP/RFQ entirely because they determine the prospective customer is not their ideal customer or because the expectations of the request are misaligned with their capabilities.
My first exposure to “tenders” was back in the late 1980’s when I participated in a closed technology consulting services tender to a governmental tourism department. Naively, I was pretty confident about my abilities and those of my partners, in being able to deliver a comprehensive and coherent service that was perfectly framed in the requirements document. What I hadn’t considered, was who I would be competing with and their approach to providing proof of past successes. The price of effectively the offer, it seemed, would not be the only selection criteria.
So this then, brings us to the competitive selection process. It isn’t always a bidding war. In cases where the RFP/RFQ responses reveal that multiple technology solutions demonstrate parity with similar functionality and capabilities, the decision-making process is more intricate. Identifying the subtle differentiators among competing products needs careful consideration in selecting the best-fit solution.
The hard comparisons
Exposure to the UI is often had, by asking for live or recorded demonstrations. Though these are often delivered in a heavily managed way, the User experience (UX) is a significant factor that can sway the decision. A technology solution that appears to be intuitive, user-friendly, and purportedly requires minimal training for employees will likely be preferred over one that is complex and seemingly difficult to navigate. Usability assessments, user testing, and interface evaluations may be the next step and may help gauge the product’s intuitiveness and its potential impact on productivity. These typically occur, when using evaluations, asking for Proof of Concept (POC) demos, or the like.
The ability of a technology solution to seamlessly integrate with the organization’s existing systems and infrastructure can significantly impact its perceived value. It is one thing to say or respond that the integration is there, it is another thing when the integration is physically demonstrated or proven. Compatibility and interoperability are often essential considerations, they can reduce implementation complexity and overall cost. Organizations assess the extent of existing integrations, the ease of connecting with a new solution, and the potential for future integration needs or promises.
Scale is important too. Organizations often prioritize solutions that can grow and adapt to future needs and growth. A technology solution that can accommodate expansion and changes in requirements ultimately offers longer-term value. Scalability assessments involve evaluating how the solution performs under various load scenarios, such as increased user counts or expanded datasets or integrations are also something which may be tested or require detailed testing evidence.
The level of support and maintenance offered by the vendor can heavily influence the decision. A responsive support team and well-defined service level agreements (SLAs) are often deemed critical, especially in complex technology implementations. The availability of a 24/7 help desk, an online knowledge base, and timely issue resolution are aspects that can significantly impact the organization’s overall satisfaction with the vendor.
Softer comparators
Execs often don’t like to hear that people count. The preferred perspective is that the solution and the company’s reputation stand for itself. The idea that buying IBM for example was a safe technology bet, was something that was echoed in many IT dept halls in the 1960s, ’70s and ’80s even though it was never actually the company’s tagline. Though IBM never used the phrase in advertising, some strap lines did allude to it: “At IBM, we sell a good night’s sleep.” per Origin.
“It was thought up by someone in corporate America cleverly commentating on the danger to one’s job when selecting pricey software that… didn’t quite pan out.” opines Cole Shafer on Honey Copy.
Another view might say, well if you’re spending your time looking for something that is likely tried and tested, go with one of the “big boys” as opposed to a Wang, RCA, Data General, Digital Equipment Corporation (DEC), Control Data Corporation (CDC), Scientific Data Systems (SDS/Xerox) who, like the below mid-1980s advert from Wang might use the phrase to explain away a failed tender to their own execs.
.
But consider yet another perspective the one where the prospective customer chases the underdog because they feel that they have a greater ability to influence the design, feel, and roadmap of a particular product because they feel that they might have leverage. This may seem counter intuitive, but it is definitely a potential factor to consider. Assuming that because you’re the underdog you don’t have a chance, may be the wrong assumption to make.
When technology solutions reach parity in functionality and capabilities, various other factors play pivotal roles in the decision-making process and some of these are unpalatably soft for execs to stomach.
The Midas touch
The personality, knowledge, expertise, and responsiveness of the customer account manager and presales consultant can leave a lasting impact on the prospect.
A strong relationship with the vendor’s representatives instills confidence in the partnership and can lead to better collaboration throughout the deal brokerage, the implementation, and beyond.
Organizations will often indirectly look for account managers and consultants who take the time to understand their unique challenges and propose appropriately tailored solutions or product positioning that align with their specific needs.
An empathetic approach that focuses on building trust and addressing concerns can ultimately be a deal-maker since it fosters a more positive and productive relationship between vendor and client.
The engagement process proposed by the vendor, including project management methodologies and communication practices, is also crucial.
A well-structured engagement plan can lead to smoother implementations and successful outcomes, basically describing to the buyer, how you plan to present your organization and offerings to the prospect.
Organizations often evaluate the vendor’s approach to project management, by looking at how they manage the opportunity as a project, including the allocation of presales and supplementary support resources, communication frequency, and risk mitigation strategies.
Effective project management processes even in the sales cycle, help ensure that the pitch stays on track and that potential derailing issues are addressed promptly.
While pricing is always going to be a factor, it becomes even more crucial when technologies are at parity. Organizations may consider other factors such as the spread of costs, including upfront costs, ongoing maintenance expenses, licensing models, and potential hidden costs.
An unseen (by the vendor) comprehensive cost-benefit analysis may be conducted, considering both short-term and long-term financial implications. This means, that as the seller you must strike a balance between budget constraints and the perceived value of the technology solution.
It’s hard to be bigger than you are, but the size and reputation of the vendor’s company is influential.
Established companies with a strong track record of successful implementations and a significant customer base may be perceived as more reliable. However, smaller companies with niche expertise may also offer unique advantages, such as personalized service and a higher level of attention to individual client needs. Organizations must evaluate their risk tolerance and assess the potential benefits and drawbacks associated with companies of different sizes.
A way to compensate for size is with physical and written, and telephonic references. Customer references and case studies provide valuable insights into real-world experiences with the solution.
Organizations often seek feedback from existing customers to gauge satisfaction and success rates. Engaging in conversations with references allows the organization to ask specific questions related to its unique requirements, implementation challenges, and the vendor’s responsiveness to their needs. Additionally, case studies that showcase successful implementations in organizations with similar profiles offer valuable validation of the technology’s suitability.
Pudding time
Some technology vendors simply won’t entertain a PoC at all. The rationale for not offering a PoC is legion. Among them, the endless PoC with seemingly infinite levels of scope creep, poorly articulated success criteria, costs to serve, deciding how to support etc
A Proof of Concept (PoC) can be a powerful tool in the decision-making process. It allows the organization to evaluate how the technology performs against specific expectations and for specific use cases before making a final commitment.
A well-designed PoC should focus on validating critical aspects, such as performance, security, scalability, and integration capabilities but success hinges on organizations working closely with the vendor to define clear success criteria for the PoC to ensure that the evaluation process remains objective and aligned with their objectives.
Communication
This is really more, of the softer stuff, but ultimately it needs to be acknowledged that in technology decision-making, the human element plays a significant enough role in shaping perceptions and building trust that it is not uncommon to see vendors roll in a larger cast of characters as the opportunity manager see the deal teeter on a buy/abandon decision.
In my days as a customer and working on the vendor side, the “suits” would often arrive when special encouragement or reassurances were needed. I was always amused for example, by the chocolate brown suits that the IBM account execs would often wear, and the shiny blue suits that the guys from SAP would often be seen wearing. I don’t remember the ones the Compaq and HP guys wore but everyone was in a suit!
The most competent account managers and presales consultants who genuinely understand the organization’s pain points and challenges pitch and propose solutions that resonate resoundingly with the organization’s objectives. They’re active listeners and respond when asked, as opposed to those who might listen to respond. Assuaging an organization’s concerns, asking clarifying questions, and demonstrating empathy for the challenges the organization faces all help in developing a deeper understanding of the organization’s unique context.
Effective communication and responsiveness build confidence in the vendor’s ability to address any concerns or issues that may arise. Timely responses to queries and proactive communication foster a sense of partnership and reliability. Candid and transparent communication about timelines, milestones, and potential risks helps manage expectations and allows the organization to plan accordingly.
Big purchases are all about partnerships rather than just transactions. A long-term commitment from the vendor fosters a sense of security and suggests a culture of collaboration. Vendors that demonstrate a vested interest in the organization’s success are perceived to be more likely to provide good ongoing support, upgrades, and enhancements, and commit to collaborate on future initiatives.
To foster a long-term partnership, organizations will always seek out vendors who prioritize customer success and demonstrate a commitment to continuous improvement.
What’s your perspective, or do you think it is all a perfect confluence of timing, product, price place, and people – essentially luck?
Every company today is a data company. Organizations worldwide are striving to derive business insights from the enormous amounts of data that are captured and stored to measure and improve business performance. McKinsey found that insight-driven companies report EBITDA (earnings before interest, taxes, depreciation, and amortization) increases of up to 25%. According to Forrester, organizations that use data […]
Today, one of the most popular techniques to achieve high levels of performance and reliability in your software applications is by leveraging the power of microservices architecture. This architectural style breaks down a monolithic application into smaller, more manageable services that can be independently developed, deployed, and scaled. While this approach offers numerous benefits, it […]
The campaign for responsible use of Artificial Intelligence (AI) has grown like a massive wildfire, and the magnitude of theproblem is growingfaster than authorities can keep up with. Agencies around the world are working to make sense of it all and provide practical solutions for change. For global business leaders, this means staying informed about compliance, ethical standards, and innovation surrounding the ethical use of AI. To date, here’s the state of AI regulation and legislation around the globe:
While the following is not a comprehensive list, it shows the distance that needs to be travelled to adequately regulating AI.
United States
In the U.S., progress toward regulating AI is well underway. The Federal Trade Commission has been working to join the campaign, starting with appending responsible AI to current laws. The burden of change is placed on business leaders to hold themselves accountable for mitigating bias. In April 2020, the FTC published a blog covering U.S. AI regulationto warn and guide businesses about the misuse of AI.
“The use of AI tools should be transparent, explainable, fair and empirically sound,” Andrew Smith, Bureau of Consumer Protection at the FTC, stated. In the release, Smith highlighted some important points for businesses using AI to remember:
Transparency in collection and use of data
Explain decision-making to consumers
Fair decision-making
Robust, empirically-sound data and modeling
Accountability for compliance, ethics, fairness and nondiscrimination
Thus far, they’ve accomplished regulating the equitable use of AI under:
The Fair Credit Reporting Act (FCRA): Biased algorithms used in housing, employment, insurance, and credit decisions are banned.
The FTC Act (FTCA): Bans the use of racially discriminatory bias in AI commercial use.
The Equal Credit Opportunity Act (ECOA): Prohibits discrimination in credit decision-making based on race, color, religion, nationality, sex, marital status, age, or the use of public assistance. Discriminatory AI is banned against “protected classes.”
In 2022, the Equal Employment Opportunity Commission (EEOC) released technical assistance guidance for algorithmic bias in employment decisions, based on the provisions under the Americans with Disabilities Act (ADA). Charlotte Burrows, Chair of the EEOC, reported that more than 80% of all employers and more than 90% of Fortune 500 companies are using such technology. Although there aren’t any federal laws that specifically target use of AI, they serve as the foundation for future legislation and regulations.
Europe
Europe has been working on regulating the commercial use of technology since 2018. The General Data Protection Regulation (GDPR) is a resource for achieving and maintaining compliance with Europe’s laws regarding the responsible use of AI. There has been much debate amongst executives and regulators regarding the European Union’s enactment of a comprehensive set of rules for governing artificial intelligence. Executives are arguing that the rules will make it difficult to contend with international competitors.
“Europe is the first regional bloc to significantly attempt to regulate AI, which is a huge challenge considering the wide range of systems that the broad term ‘AI’ can cover,” said Sarah Chander, senior policy adviser at digital rights group EDRi.
China
In 2017, the Chinese State Council released the Next Generation Artificial Intelligence Development Plan as a set of guidelines surrounding the use of specific AI applications. The release was regarding currently active provisions on the management of algorithmic recommendations of Internet information services and the management of deep synthesis of Internet information services, which is still being drafted.
In May 2023, China’s Cyberspace Administration (CAC) drafted the Administrative Measures for Generative Artificial Intelligence Services. It requires a “safety assessment” for companies desiring to develop new AI products before they can go to market. It also mentions the use of truthful, accurate data, free of discriminatory algorithms. It focuses on prevention as the major first step for responsible AI.
Brazil
In December 2022, Brazilian Senators released a report containing studies and a draft of a regulation relating to responsible AI governance. It serves to inform future regulations that Brazil’s Senate is planning. The focal point of the regulation was the presentation of three central pillars:
Guaranteeing the rights of people AI affects
Classification of risk levels
Predicting Governance Measures
Japan
In March 2019, Japan’s Integrated Innovation Strategy Promotion Council created the Social Principles of Human-Human-Centric AI. The two-part provision is meant to address a myriad of social issues that have come with AI innovation. One part established seven social principles to govern the public and private use of AI:
Human-centricity
Education/literacy
Data protection
Ensuring safety
Fair competition
Fairness
Accountability & Transparency
Innovation
The other part, which expounds on the 2019 provision, targets AI developers and the companies that employ them. The AI Utilisation Guidelines are meant to be an instruction manual for AI developers and companies to develop their own governance strategy. There’s also the 2021 provision, Governance Guidelines for Implementation of AI Principles, which features hypothetical examples of AI applications for them to review. While none of these regulations are legally binding, they are Japan’s first step in starting the race to regulating AI.
Canada
In June 2022, Canada’s federal government released theDigital Charter Implementation Act. This contained Canada’s first piece of legislation to strengthen the country’s efforts to mitigate bias. The charter included the Artificial Intelligence and Data Act, which regulates international and interprovincial trade in AI. It requires that developers responsibly ensure to mitigate risk and bias. Public disclosure requirements and prohibitions on harmful use are also included. The charter is preliminary to moving toward officially enacting legislation regarding AI in Canada.
India
Currently, there are no official regulatory requirements in India regarding the responsible use of AI. The Indian Commission NITI Aayog has released working research papers being used to begin to address the issues. The first installment of the paper, Towards Responsible #AIforAll, discusses the potential of AI for society at large and recommendations surrounding AI adoption in the public and private sectors. The next part, an Approach Document for India, established principles for responsible AI, the economic potential of AI, supporting large-scale adoption, and establishing and instilling public trust. The final paper, Adopting the Framework: A Use Case Approach on Facial Recognition Technology, is meant to be a “benchmark for future AI design, development, and deployment in India.”
Switzerland
There are currently no specific regulations that govern the responsible use of AI. Already enacted laws are being used to inform cases as they present themselves. For example, the General Equal Treatment Act, their product liability and general civil laws address prevention of bias in the public and private sectors.
The Future of a Global Approach
To limit or completely eradicate AI bias, there needs to be a communal effort and commitment to accuracy, trust, and compliance. Business leaders and developers should target preventive, corrective and measures for transparency, accuracy, and accountability when employing AI. Regulators must also do their due diligence in providing comprehensive, appropriate, and timely legislation that applies to the present and will be relevant in the future.
In this digital era, the emergence of AI has opened up avenues for a remarkable companion in combatting mental health issues such as depression. Ready yourself as we investigate how AI chatbots are rewriting the rules of mental health care and providing new hope and advanced strategies for managing depression. One of the most advantageous […]
Driven by the desire for organizations to get better business insights, data systems are becoming more specialized, and data stacks are increasing in complexity. As companies continue their quest toward data-driven operations, they must balance speed and cost. This is why we recently engaged with GigaOm Research to conduct a TPC-H Benchmark Test against Snowflake and BigQuery – the results were clear, the Actian cloud data platform offers superior performance at a fraction of the cost of these competitors.
The Actian platform’s operational data warehouse is designed to support real-time data analytics so customers can maintain a competitive advantage. The TPC-H benchmark consists of a series of ad-hoc analytical queries that involve complex joins, aggregations, and sorting operations. These queries represent common decision support tasks to generate sales reports, analyze trends, and perform advanced data analytics. In today’s rapidly changing business climate, there is no room for delays when it comes to accessing data to support business decisions.
Our data analytics engine ensures that the warehouse capability in the Actian platform delivers on the promise of performance without runaway costs. The GigaOm field test, informed by TPC-H spec validation queries, highlights the price and performance ratio and cost-effectiveness of the Actian platform, providing independent validation of the Actian data platform in terms of both performance and cost.
The Results
In the GigaOm benchmark, the Actian Data Platform outperformed both Snowflake and BigQuery in 20 of the 22 queries, clearly illustrating Actian’s powerful decision support capabilities. Leveraging decades of data management experience, the Actian platform provides data warehouse technology that uses in-memory computing along with optimized data storage, vector processing, and query execution that exploits powerful CPU features. These capabilities significantly improve the speed and efficiency of real-time analytics.
The benchmark results reveal query execution and price efficiencies that outperform competitor solutions, lowering the total cost of ownership without sacrificing speed. Overall, the Actian platform delivered query results that were 3x faster than Snowflake and 9x faster than BigQuery. Performance improved with additional users, highlighting the platform’s ability to scale with concurrency to meet the demands of all business users.
In terms of cost, the GigaOm field tests further prove the value of the Actian data platform over the competition. Snowflake’s costs were nearly 4x higher than Actian’s, and BigQuery ranged from 11x to 16x more expensive based on concurrency.
To counter current economic pressures such as inflation, scarce talent, and supply constraints, 80% of companies are increasing investments in digital technology, according to a report by Gartner. Driving these initiatives are IT leaders, who – depending on the industry they work in – are increasingly doing so in operational environments like factories, production sites, and even […]
I work with a lot of organizations that do not (yet) have a person in the role of Chief Data Officer (CDO) or Chief Data and Analytics Officer (CDAO). I also work with a significant number of people in these organizations who have the desire to be elevated into one of these roles. The latter […]
The world revolves around data. Just about any important decision organizations make depends on the data they have. Data helps organizations mitigate risks, plan marketing campaigns, devise research-backed strategies, and tons more. But a resource this valuable has rather humble, or jumbled, beginnings. What is Raw Data? Raw data is the primary form of data. […]
My company led a Data Management Assessment for a government agency this spring, using the Data Management Capability Assessment Model (DCAM), and delivered the final report at the end of May. In my last column comparing the DCAM to the DMM, I said that I’d share my observations with you so that you can get […]
It is no secret that cloud migration and transformation helps your business attain desired growth, cost savings, agility, and profitability. With these, your team is able address customer needs faster, monitor app performance, and scale applications according to demand. App migration can be time-consuming, and you cannot afford to mess with the process. And a […]
Businesses today rely on real-time big data analytics to handle the vast and complex clusters of datasets. Here’s the state of big data today: The forecasted market value of big data will reach $650 billion by 2029. From 2010 to 2020, there has been a 5000% growth in the quantity of data created, captured, and […]
The recently published report by Research Nester, Global Data Mining Tool Market: Global Demand Analysis & Opportunity Outlook 2027, delivers detailed overview of the global data mining tool market in terms of market segmentation by service type, function type, industry type, deployment type, and region. The report also encompasses the industry growth drivers, restraints, supply […]
In the digital age, where technology permeates almost every aspect of our lives, the protection of our online systems and sensitive data has become paramount. Passwords are the first line of defense against unauthorized access and cyber threats. Unfortunately, many users still underestimate the significance of regular password changes, often opting for convenience over security.
The threat landscape in the cybersecurity realm is continuously evolving, with hackers employing sophisticated techniques to exploit vulnerabilities in online systems. Cybercriminals often utilize brute force attacks, dictionary attacks, and credential stuffing to gain unauthorized access to accounts. Regular password changes are crucial to staying ahead of these threats and reducing the window of opportunity for attackers to compromise accounts.
Data breaches have become distressingly common, exposing millions of user credentials to cybercriminals. One of the main reasons behind these breaches is the reuse of passwords across multiple accounts. Regularly changing passwords minimizes the impact of a data breach, as even if one account is compromised, the attacker will have limited access and time to exploit other accounts associated with the same password.
For individuals, regular password changes play a pivotal role in safeguarding personal accounts. Online banking, social media profiles, email accounts, and e-commerce platforms contain a wealth of sensitive information. Regularly updating passwords ensures that even if someone gains unauthorized access to an account, their window of opportunity to misuse that information is limited.
In a corporate setting, password security is of utmost importance to protect sensitive business data and maintain employee productivity. Regular password changes are a fundamental aspect of any robust cybersecurity policy. They act as a safety net against insider threats, disgruntled employees, or unauthorized personnel attempting to infiltrate the organization’s systems.
In various industries, businesses are bound by strict compliance regulations and legal requirements related to data protection. Regular password changes are often mandated by these regulations to maintain a certain level of security and reduce the risk of data breaches. Non-compliance can result in severe penalties and reputational damage.
Multifactor authentication (MFA) is an additional layer of security that supplements passwords by requiring users to provide multiple forms of identification. While MFA significantly strengthens security, passwords remain an integral part of the authentication process. Regular password changes complement MFA by ensuring that the primary authentication method remains up-to-date and secure.
By encouraging regular password changes, organizations can foster a culture of cybersecurity awareness among employees. It prompts individuals to think about their password choices, encouraging the use of strong, unique passwords for each account. This heightened awareness can also extend to personal online habits, benefiting users beyond their workplace.
In scenarios where users inadvertently share their passwords or forget to log out of accounts on shared devices, regular password changes act as a safety mechanism. If unauthorized individuals gain access to a password, it becomes obsolete after a short period, reducing the potential damage caused by unauthorized access.
Sometimes, users may not be aware that their accounts have been compromised until it’s too late. Regular password changes can serve as an early warning system, as sudden login attempts or suspicious activities on an account can indicate potential unauthorized access. This prompts users to take immediate action and report any suspicious behavior.
In organizations that prioritize regular password changes, employees are more likely to adopt other security best practices. A security-first culture fosters an environment where individuals actively seek to protect the organization and its data, making the entire system more resilient to cyber threats.
Regular password changes are an indispensable element of a robust cybersecurity strategy for both individuals and organizations.
By staying ahead of the ever-evolving threat landscape, preventing data breaches, enhancing workplace security, and fostering cybersecurity awareness, the simple act of updating passwords plays a significant role in protecting sensitive information and maintaining online safety.
As technology continues to advance, prioritizing the importance of regular password changes remains a fundamental pillar in our ongoing battle against cyber threats.
According to Gartner, 80% of executives believe organizations can apply AI to any business decision. However, an AI-based decision is only as effective as the training data on which the AI model is built. The most effective, advanced solutions focus on domain-specific AI and ensure training data used in AI models is highly accurate and reliable. […]