WO2013025553A2 - Data volume management - Google Patents

Data volume management Download PDF

Info

Publication number
WO2013025553A2
WO2013025553A2 PCT/US2012/050452 US2012050452W WO2013025553A2 WO 2013025553 A2 WO2013025553 A2 WO 2013025553A2 US 2012050452 W US2012050452 W US 2012050452W WO 2013025553 A2 WO2013025553 A2 WO 2013025553A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
index store
project
user
frontend user
Prior art date
Application number
PCT/US2012/050452
Other languages
French (fr)
Other versions
WO2013025553A3 (en
Inventor
Robin Kumar DAS
Ledio Ago
Declan Gerard SHANAGHY
Gaurav Gupta
Original Assignee
Splunk Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Splunk Inc. filed Critical Splunk Inc.
Publication of WO2013025553A2 publication Critical patent/WO2013025553A2/en
Publication of WO2013025553A3 publication Critical patent/WO2013025553A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/10Network architectures or network communication protocols for network security for controlling access to devices or network resources
    • H04L63/105Multiple levels of security
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/13File access structures, e.g. distributed indices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/13File access structures, e.g. distributed indices
    • G06F16/134Distributed indices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/16File or folder operations, e.g. details of user interfaces specifically adapted to file systems
    • G06F16/162Delete operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/21Design, administration or maintenance of databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2228Indexing structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2228Indexing structures
    • G06F16/2272Management thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/901Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/955Retrieval from the web using information identifiers, e.g. uniform resource locators [URL]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/958Organisation or management of web site content, e.g. publishing, maintaining pages or automatic linking
    • G06F16/986Document structures and storage, e.g. HTML extensions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
    • G06F9/4881Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
    • G06F9/4887Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues involving deadlines, e.g. rate based, periodic
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system
    • G06F9/5088Techniques for rebalancing the load in a distributed system involving task migration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/08Payment architectures
    • G06Q20/14Payment architectures specially adapted for billing systems
    • G06Q20/145Payments according to the detected use or quantity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/14Charging, metering or billing arrangements for data wireline or wireless communications
    • H04L12/1432Metric aspects
    • H04L12/1435Metric aspects volume-based
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/10Network architectures or network communication protocols for network security for controlling access to devices or network resources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]

Definitions

  • the invention is directed to providing services for use by a customer, and more particularly, to providing search services to the customer over a network with a cloud-based system.
  • the volume of data processed by a node in a small business can be a magnitude less than the volume of data processed by a node in a large corporation.
  • the amount of data to be indexed for searching and the need for speed to search the indexed data can vary.
  • the type of data processed by large and small business can vary widely, such as sending and receiving messages, creating and storing documents, hosting web sites, database searches, facilitating online transactions, and the like.
  • the expense of developing and maintaining software that can handle the type and volume of data processed by a large corporation can be substantially greater than the effort expended to do somewhat the same for a small business. Consequently, cloud-based systems that provide software as a service are increasingly popular with different sizes of businesses having different volumes of data and types of data.
  • FIGURE 1 illustrates a system environment in which various embodiments may be implemented
  • FIGURE 2A shows a rack of blade servers that may be included in various embodiments
  • FIGURE 2B illustrates an embodiment of a blade server that may be included in a rack of blade servers such as that shown in FIGURE 2 A;
  • FIGURE 3 shows a mobile device that may be included in various embodiments
  • FIGURE 4 illustrates a network device that may be included in various embodiments
  • FIGURE 5 illustrates a system diagram showing an embodiment of a system that may be utilized to implement embodiments of processing, searching and managing data volumes;
  • FIGURE 6 illustrates a system diagram showing an alternative embodiment of a system that may be utilized to implement embodiments of processing, searching and managing data volumes;
  • FIGURE 7 illustrates a logical flow diagram generally showing one embodiment of an overview process for initiating a project that employs embodiments of the system diagrams shown in FIGURES 5 and/or 6;
  • FIGURE 8 illustrates a logical flow diagram generally showing one embodiment of a process for enabling access to a determined volume of data and restricting access to incoming data that exceeds the determined volume;
  • FIGURE 9 illustrates a logical flow diagram generally showing one embodiment of a process for managing an overflow index store
  • FIGURE 10 illustrates a logical flow diagram generally showing one embodiment of a process for enabling a project owner to invite another user into the project
  • FIGURE 11 illustrates a logical flow diagram generally showing one embodiment of an overview process for elastically scaling a project
  • FIGURE 12 illustrates a logical flow diagram generally showing one embodiment of a process for receiving resource allocation direction from an operator manipulating a user interface control
  • FIGURE 13 illustrates a logical flow diagram generally showing one embodiment of a process for increasing a project's resources as part of elastic scaling
  • FIGURE 14 depicts a logical flow diagram generally showing one embodiment of a process for responding to a requested decrease in project resources
  • FIGURE 15 illustrates a logical flow diagram generally showing one embodiment of process for queuing resource requests
  • FIGURE 16 depicts a logical flow diagram generally showing one embodiment of process for elastically scaling based on index storage size
  • FIGURE 17 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on indexing performance
  • FIGURE 18 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on an event time window
  • FIGURE 19 illustrates a logical flow diagram generally showing one embodiment of a process for changing a number of indexers associated with an index store
  • FIGURE 20 shows a logical flow diagram generally showing one embodiment of a process to recall archived data for re -indexing
  • FIGURE 21 illustrates a logical flow diagram generally showing one embodiment of a process for transitioning to scaled index stores
  • FIGURE 22 illustrates a logical flow diagram generally showing one embodiment of an input proxy operation
  • FIGURE 23 illustrates a non-exhaustive example of a use case of an embodiment of user-interface controls for managing and processing data volumes.
  • Frontend User may represent users at the outer cloud-based edge of some of the described embodiments. This outer edge may often be described herein as the Web Application Layer, or Cloud layer.
  • Core User may represent abstract users that are within the middle of the layers of described embodiments. These middle layers may often be described herein as Search Head layers, or Core layers.
  • Apps as used herein may collectively represent software programs and processes that may implement various features and processing for the various embodiments, some of which are described herein. Apps may give users insights into their projects with dashboards, reports, data inputs, and saved searches that may work in the project environment from the moment they are installed. Apps may include new views and dashboards that may completely reconfigure the way a project looks. Or, they may be as complex as an entirely new program. In some embodiments, Apps may also represent index resources. "Index”, “Indexer”, “Indexing”, and “Index Storage” as used herein may represent elements of described embodiments that may index and store data and events. Indexers may collect, parse, and store data to facilitate fast and accurate information retrieval.
  • Index design may incorporate interdisciplinary concepts from linguistics, cognitive psychology, mathematics, informatics, physics, and computer science. Also, Indexes may reside in flat files in a data store on a file system. Index files may be managed to facilitate flexible searching and fast data retrieval, eventually archiving them according to a configurable schedule. During indexing, incoming raw data may be processed to enable fast search and analysis, the results of which may be stored in an index. As part of the indexing process, the indexer may add knowledge to the data in various ways, including by: separating a data stream into individual, searchable events; creating or identifying timestamps; extracting fields such as host, source, and source type;
  • index store may store data indexed by one or more indexers.
  • indexer may include indexer severs.
  • Roles may represent definitions of the access permissions and authorizations that enable a Core User to execute various features, Apps, or access data, in the various embodiments. Roles may define how a Core User may interact with Apps, Indexes, Indexers, and the like. "Dimension” as used herein when describing elastic scaling, may refer to the particular kind of project resources that may be scaled, e.g., index storage, processing performance, time- window, price, network bandwidth, and the like.
  • Embodiments may be arranged and configured to define dimensions differently, e.g., one embodiment may be arranged to define an elastic scaling dimension that represents network latency, where another embodiment may be arranged and configured to combine CPU quota, network latency, network bandwidth into a single dimension representing processing performance.
  • an elastic scaling dimension that represents network latency
  • another embodiment may be arranged and configured to combine CPU quota, network latency, network bandwidth into a single dimension representing processing performance.
  • wi 11 : ⁇ L ⁇ that there may be many combinations of resources that may arranged into elastic scaling dimensions.
  • Multi-tenant cloud-based data storage volumes may be built to support manual and rule-based elastic scaling.
  • Elastic scaling may occur in one or more dimensions, such as scaling of storage, scaling of processing performance, or time windows of received data, and the like.
  • Elastic scaling may occur in one or more dimensions singly or in combination. If elastic scaling occurs, the appropriate objects, processes, memory, and storage may be scaled to meet the request resource increase or decrease. There may be numerous dependent or supporting systems that may need to scale in response to an elastic scaling request. For example, a request to elastically scale processing performance may require one or more elements of the system to be scaled in addition to the CPU quota, such as, disk quota, high local cache quota, and the like.
  • various embodiments are directed towards a system and method for a cloud-based front end that may abstract and enable access to the underlying cloud-hosted elements and objects that may be part of a multi-tenant application, such as a search application.
  • Search objects may be employed to access indexed objects.
  • one or more projects can be elastically scaled for a user to provide resources that may meet the specific needs of each project.
  • At least one project for at least one frontend user may be initialized.
  • the frontend user may be enabled to interact with the project through at least one application.
  • Each project may have a corresponding index store, which may include one or more indexer servers.
  • a data volume limit may be established for the index store. The data volume limit may be based on a selection by the frontend user.
  • Data provided for the project may be indexed. In some embodiments, if an amount of indexed data in the index store is less than the data volume limit, then the indexed data may be stored in the index store. In other embodiments, if the amount of indexed data in the index store at least equals the data volume limit, then the indexed data may be stored in an overflow index store.
  • the indexed data in the index store may be accessible to the frontend user and the indexed data in the overflow index store may be inaccessible to the frontend user (e.g., the frontend user may be prevented from accessing the data unless the user increases the data volume limit, such as by purchasing additional data storage).
  • the data volume limit is increased, then at least a portion of indexed data in the overflow index store may be provided to the index store, such that the provided indexed data is accessible in the index store to the frontend user. In at least one embodiment, the data volume limit may be increased based on another selection by the frontend user.
  • the user may be enabled to elastically scale one or more computing resources associated with a project.
  • a resource allocation may be determined for indexing data provided for a project for each of a plurality of resources.
  • the resources may include at least one indexer server instance that indexes the provided data, a data store (i.e., storage capacity) that stores the indexed data, an event time window, or the like.
  • a request may be provided from a user to scale at least one of the plurality of resources.
  • the user may be enabled to utilize a slider control (or other user interface) to manually adjust one or more computing resources.
  • the resource allocation for at least one of the plurality of resources may be adjusted based on the request.
  • At least one of the plurality of resources may be changed based on the adjusted resource allocation, which may be employed to subsequently index data provided for the project.
  • at least one type of additional resource may be allocated if the request indicates an increase in the plurality of resources. In other embodiments, less of at least one type of resource may be allocated if the change indicates a decrease in the plurality of resources.
  • an increase or decrease in one resource may correspond to an increase or decrease in another resource.
  • a number of indexers may be increased or decreased to accommodate the change in resources.
  • decreasing a number of indexers may include marking an indexer as read-only and increasing a number of indexers may include marking a read-only indexer as read/write and/or adding a new indexer.
  • changing resources may include changing a number of indexer servers employed to index data for the project, changing an amount of storage for the data store to store data for the project, changing an amount of memory employed to process the indexing of data for the project, changing an amount of processors employed to process the indexing of data for the project, or the like.
  • FIGURE 1 shows components of an environment in which various embodiments may be practiced. Not all of the components may be required to practice the various embodiments, and variations in the arrangement and type of the components may be made without departing from the spirit or scope of the various embodiments.
  • cloud network 102 enables one or more network services for a user based on the operation of corresponding arrangements 104 and 106 of virtually any type of networked computing device.
  • the networked computing devices may include server network device 112, host network device 114, enclosure of blade servers 110, enclosure of server computers 116, super computer network device 118, and the like.
  • one or more mobile devices may be included in cloud network 102 in one or more arrangements to provide one or more network services to a user. Also, these arrangements of networked computing devices may or may not be mutually exclusive of each other.
  • the user may employ a plurality of virtually any type of wired or wireless networked computing devices to communicate with cloud network 102 and access at least one of the network services enabled by one or more of arrangements 104 and 106.
  • These networked computing devices may include tablet mobile device 122, handheld mobile device 124, wearable mobile device 126, desktop network device 120, and the like.
  • the user may also employ notebook computers, desktop computers,
  • microprocessor-based or programmable consumer electronics network appliances, mobile telephones, smart telephones, pagers, radio frequency (RF) devices, infrared (IR) devices, Personal Digital Assistants (PDAs), televisions, integrated devices combining at least one of the preceding devices, and the like.
  • RF radio frequency
  • IR infrared
  • PDA Personal Digital Assistants
  • mobile devices may include virtually any substantially portable networked computing device capable of communicating over a wired, wireless, or some combination of wired and wireless network.
  • network 102 may employ virtually any form of communication technology and topology.
  • network 102 can include local area networks Personal Area Networks (PANs), (LANs), Campus Area Networks (CANs), Metropolitan Area Networks (MANs) Wide Area Networks (WANs), direct communication connections, and the like, or any combination thereof.
  • PANs Personal Area Networks
  • LANs Local Area Networks
  • CANs Campus Area Networks
  • MANs Metropolitan Area Networks
  • WANs Wide Area Networks
  • direct communication connections and the like, or any combination thereof.
  • a router acts as a link between LANs, enabling messages to be sent from one to another.
  • communication links within networks may include virtually any type of link, e.g., twisted wire pair lines, optical fibers, open air lasers or coaxial cable, plain old telephone service (POTS), wave guides, acoustic, full or fractional dedicated digital communication lines including Tl , T2, T3, and T4, and/or other carrier and other wired media and wireless media.
  • POTS plain old telephone service
  • carrier mechanisms may include E-carriers, Integrated Services
  • ISDNs Digital Networks
  • USB universal serial bus
  • Firewire ports Firewire ports
  • Thunderbolt ports Digital Subscriber Lines (DSLs)
  • wireless links including satellite links, or other
  • communications links known to those skilled in the art.
  • these communication links may further employ any of a variety of digital signaling technologies, including without limit, for example, DS-0, DS-1, DS-2, DS-3, DS-4, OC-3, OC-12, OC-48, or the like.
  • remotely located computing devices could be remotely connected to networks via a modem and a temporary communication link.
  • network 102 may include virtually any communication technology by which information may travel between computing devices.
  • the communicated information may include virtually any kind of information including, but not limited to processor-readable instructions, data structures, program modules, applications, raw data, control data, archived data, video data, voice data, image data, text data, and the like.
  • Network 102 may be partially or entirely embodied by one or more wireless networks.
  • a wireless network may include any of a variety of wireless sub-networks that may further overlay stand-alone ad-hoc networks, and the like. Such sub-networks may include mesh networks, Wireless LAN (WLAN) networks, Wireless Router (WR) mesh, cellular networks, pico networks, PANs, Open Air Laser networks, Microwave networks, and the like.
  • Network 102 may further include an autonomous system of intermediate network devices such as terminals, gateways, routers, switches, firewalls, load balancers, and the like, which are coupled to wired and/or wireless communication links. These autonomous devices may be operable to move freely and randomly and organize themselves arbitrarily, such that the topology of network 102 may change rapidly.
  • Network 102 may further employ a plurality of wired and wireless access technologies, e.g., 2nd (2G), 3rd (3G), 4th (4G), 5 th (5G) generation wireless access technologies, and the like, for mobile devices.
  • These wired and wireless access technologies may also include Global System for Mobile communication (GSM), General Packet Radio Services (GPRS), Enhanced Data GSM Environment (EDGE), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution Advanced (LTE), Universal Mobile Telecommunications System (UMTS), Orthogonal frequency-division multiplexing (OFDM), Wideband Code Division Multiple Access (W-CDMA), Code Division Multiple Access 2000 (CDMA2000), Evolution-Data Optimized (EV-DO), High-Speed Downlink Packet Access (HSDPA), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), ultra wide band (UWB), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), any portion of the Open Systems Inter
  • FIGURE 2A shows one embodiment of an enclosure of blade servers 200, which are also illustrated in FIGURE 1.
  • Enclosure of blade servers 200 may include many more or fewer components than those shown in FIGURE 2A. However, the components shown are sufficient to disclose an illustrative embodiment.
  • a blade server is a stripped down server computing device with a modular design optimized to minimize the use of physical space and energy.
  • a blade enclosure can include several blade servers and provide each with power, cooling, network interfaces, input/output interfaces, and resource management.
  • an enclosure of server computers typically includes several computers that merely require a network connection and a power cord connection to operate. Each server computer often includes redundant components for power and interfaces.
  • enclosure 200 contains power supply 204, and input/output interface 206, rack logic 208, several blade servers 210, 212, 214, and 216, and backplane 202.
  • Power supply 204 provides power to each component and blade server within the enclosure.
  • the input/output interface 206 provides internal and external communication for components and blade servers within the enclosure.
  • Backplane 208 can enable passive and active communication of power, logic, input signals, and output signals for each blade server.
  • FIGURE 2B illustrates an illustrative embodiment of blade server 250, which may include many more or fewer components than those shown. As shown in FIGURE 2A, a plurality of blade servers may be included in one enclosure that shares resources provided by the enclosure to reduce size, power, and cost. Blade server 250 includes processor 252 which communicates with memory 256 via bus
  • Blade server 250 also includes input/output interface 290, processor-readable stationary storage device 292, and processor-readable removable storage device 294.
  • Input/output interface 290 can enable blade server 250 to communicate with other blade servers, mobile devices, network devices, and the like.
  • Interface 190 may provide wireless and/or wired communication links for blade server.
  • Processor-readable stationary storage device 292 may include devices such as an electromagnetic storage device (hard disk), solid state hard disk (SSD), hybrid of both an SSD and a hard disk, and the like.
  • processor-readable removable storage device 294 enables processor 252 to read non-transitive storage media for storing and accessing processor- readable instructions, modules, data structures, and other forms of data.
  • the non-transitive storage media may include Flash drives, tape media, floppy media, and the like.
  • Memory 256 may include Random Access Memory (RAM), Read-Only Memory (ROM), hybrid of RAM and ROM, and the like. As shown, memory 256 includes operating system 258 and basic input/output system (BIOS) 260 for enabling the operation of blade server 250. In various embodiments, a general-purpose operating system may be employed such as a version of UNIX, or LINUXTM, or a specialized server operating system such as Microsoft's Windows ServerTM and Apple Computer's IoS ServerTM.
  • RAM Random Access Memory
  • ROM Read-Only Memory
  • BIOS basic input/output system
  • a general-purpose operating system may be employed such as a version of UNIX, or LINUXTM, or a specialized server operating system such as Microsoft's Windows ServerTM and Apple Computer's IoS ServerTM.
  • Memory 256 further includes one or more data storage 270, which can be utilized by blade server 250 to store, among other things, applications 280 and/or other data.
  • Data stores 270 may include program code, data, algorithms, and the like, for use by processor 252 to execute and perform actions. In one embodiment, at least some of data store 270 might also be stored on another component of blade server 250, including, but not limited to, processor- readable removable storage device 294, processor-readable stationary storage device 292, or any other processor-readable storage device (not shown).
  • Data storage 270 may include, for example, raw data 272, and indexed data 274.
  • Applications 280 may include processor executable instructions which, when executed by blade server 250, transmit, receive, and/or otherwise process messages, audio, video, and enable communication with other networked computing devices.
  • Examples of application programs include database servers, file servers, calendars, transcoders, and so forth.
  • Applications 280 may include, for example, cloud search application 282, and indexing application 284.
  • Human interface components may be remotely associated with blade server 250, which can enable remote input to and/or output from blade server 250.
  • information to a display or from a keyboard can be routed through the input/output interface 290 to appropriate peripheral human interface components that are remotely located.
  • peripheral human interface components include, but are not limited to, an audio interface, a display, keypad, pointing device, touch interface, and the like.
  • FIGURE 3 shows one embodiment of mobile device 300 that may include many more or less components than those shown.
  • Mobile device 300 may represent, for example, at least one embodiment of mobile devices shown in FIGURE 1.
  • Mobile device 300 includes processor 302 in communication with memory 304 via bus 328.
  • Mobile device 300 also includes power supply 330, network interface 332, audio interface 356, display 350, keypad 352, illuminator 354, video interface 342, input/output interface 338, haptic interface 364, global positioning systems (GPS) receiver 358, Open air gesture interface 360, temperature interface 362, camera(s) 340, projector 346, pointing device interface 366, processor-readable stationary storage device 334, and processor-readable removable storage device 336.
  • Power supply 330 provides power to mobile device 300.
  • a rechargeable or non- rechargeable battery may be used to provide power.
  • the power may also be provided by an external power source, such as an AC adapter or a powered docking cradle that supplements and/or recharges the battery.
  • a gyroscope may be employed within mobile device 300 to measuring and/or maintaining an orientation of mobile device 300.
  • Mobile device 300 may optionally communicate with a base station (not shown), or directly with another computing device.
  • Network interface 332 includes circuitry for coupling mobile device 300 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the Open Systems Interconnection (OSI) model for mobile communication (GSM), code division multiple access (CDMA), time division multiple access (TDMA), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), Short Message Service (SMS), Multimedia Messaging Service (MMS), general packet radio service (GPRS), WAP, ultra wide band (UWB), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), Session Initiation Protocol/Real-time Transport Protocol (SIP/RTP), or any of a variety of other wireless communication protocols.
  • OSI Open Systems Interconnection
  • GSM Open Systems Interconnection
  • CDMA code division multiple access
  • TDMA time division multiple access
  • UDP user datagram protocol
  • Network interface 332 is sometimes known as a transceiver, transceiving device, or network interface card (NIC).
  • Audio interface 356 is arranged to produce and receive audio signals such as the sound of a human voice.
  • audio interface 356 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others and/or generate an audio
  • a microphone in audio interface 356 can also be used for input to or control of mobile device 300, e.g., using voice recognition, detecting touch based on sound, and the like.
  • Display 350 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computing device.
  • Display 350 may also include a touch interface 344 arranged to receive input from an object such as a stylus or a digit from a human hand, and may use resistive, capacitive, surface acoustic wave (SAW), infrared, radar, or other technologies to sense touch and/or gestures.
  • Projector 346 may be a remote handheld projector or an integrated projector that is capable of projecting an image on a remote wall or any other reflective object such as a remote screen.
  • Video interface 342 may be arranged to capture video images, such as a still photo, a video segment, an infrared video, or the like.
  • video interface 342 may be coupled to a digital video camera, a web-camera, or the like.
  • Video interface 342 may comprise a lens, an image sensor, and other electronics.
  • Image sensors may include a complementary metal-oxide- semiconductor (CMOS) integrated circuit, charge-coupled device (CCD), or any other integrated circuit for sensing light.
  • CMOS complementary metal-oxide- semiconductor
  • CCD charge-coupled device
  • Keypad 352 may comprise any input device arranged to receive input from a user.
  • keypad 352 may include a push button numeric dial, or a keyboard.
  • Keypad 352 may also include command buttons that are associated with selecting and sending images.
  • Illuminator 354 may provide a status indication and/or provide light. Illuminator 354 may remain active for specific periods of time or in response to events. For example, when illuminator 354 is active, it may backlight the buttons on keypad 352 and stay on while the mobile device is powered. Also, illuminator 354 may backlight these buttons in various patterns when particular actions are performed, such as dialing another mobile device.
  • Illuminator 354 may also cause light sources positioned within a transparent or translucent case of the mobile device to illuminate in response to actions.
  • Mobile device 300 also comprises input/output interface 338 for communicating with external peripheral devices or other computing devices such as other mobile devices and network devices.
  • the peripheral devices may include an audio headset, display screen glasses, remote speaker system, remote speaker and microphone system, and the like.
  • Input/output interface 338 can utilize one or more technologies, such as Universal Serial Bus (USB), Infrared, WiFi, WiMax, BluetoothTM, and the like.
  • Haptic interface 364 is arranged to provide tactile feedback to a user of the mobile device. For example, the haptic interface 364 may be employed to vibrate mobile device 300 in a particular way when another user of a computing device is calling.
  • Temperature interface 362 may be used to provide a temperature measurement input and/or a temperature changing output to a user of mobile device 300.
  • Open air gesture interface 360 may sense physical gestures of a user of mobile device 300, for example, by using single or stereo video cameras, radar, a gyroscopic sensor inside a device held or worn by the user, or the like.
  • Camera 340 may be used to track physical eye movements of a user of mobile device 300.
  • GPS transceiver 358 can determine the physical coordinates of mobile device 300 on the surface of the Earth, which typically outputs a location as latitude and longitude values. GPS transceiver 358 can also employ other geo-positioning mechanisms, including, but not limited to, triangulation, assisted GPS (AGPS), Enhanced Observed Time Difference (E-OTD), Cell Identifier (CI), Service Area Identifier (SAI), Enhanced Timing Advance (ETA), Base Station Subsystem (BSS), or the like, to further determine the physical location of mobile device 300 on the surface of the Earth. It is understood that under different conditions, GPS transceiver 358 can determine a physical location for mobile device 300.
  • AGPS assisted GPS
  • E-OTD Enhanced Observed Time Difference
  • CI Cell Identifier
  • SAI Service Area Identifier
  • ETA Enhanced Timing Advance
  • BSS Base Station Subsystem
  • mobile device 300 may, through other components, provide other information that may be employed to determine a physical location of the device, including for example, a Media Access Control (MAC) address, IP address, and the like.
  • Human interface components can be peripheral devices that are physically separate from mobile device 300, allowing for remote input and/or output to mobile device 300.
  • information routed as described here through human interface components such as display 350 or keyboard 352 can instead be routed through network interface 332 to appropriate human interface components located remotely.
  • Examples of human interface peripheral components that may be remote include, but are not limited to, audio devices, pointing devices, keypads, displays, cameras, projectors, and the like. These peripheral components may communicate over a Pico Network such as BluetoothTM, ZigbeeTM and the like.
  • a mobile device with such peripheral human interface components is a wearable computing device, which might include a remote pico projector along with one or more cameras that remotely communicate with a separately located mobile device to sense a user's gestures toward portions of an image projected by the pico projector onto a reflected surface such as a wall or the user's hand.
  • a mobile device may include a browser application that is configured to receive and to send web pages, web-based messages, graphics, text, multimedia, and the like.
  • the mobile device's browser application may employ virtually any programming language, including a wireless application protocol messages (WAP), and the like.
  • WAP wireless application protocol
  • the browser application is enabled to employ Handheld Device Markup Language (HDML), Wireless Markup Language (WML), WMLScript, JavaScript, Standard Generalized Markup Language (SGML), HyperText Markup Language (HTML), extensible Markup Language (XML), HTML5, and the like.
  • HDML Handheld Device Markup Language
  • WML Wireless Markup Language
  • WMLScript Wireless Markup Language
  • JavaScript Standard Generalized Markup Language
  • SGML Standard Generalized Markup Language
  • HTML HyperText Markup Language
  • XML extensible Markup Language
  • HTML5 HyperText Markup Language
  • Memory 304 may include Random Access Memory (RAM), Read-Only Memory (ROM), and/or other types of memory.
  • Memory 304 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Memory 304 stores a basic input/output system (BIOS) 308 for controlling low-level operation of mobile device 300.
  • BIOS basic input/output system
  • the memory also stores an operating system 306 for controlling the operation of mobile device 300.
  • this component may include a general-purpose operating system such as a version of UNIX, or LINUXTM, or a specialized mobile computer communication operating system such as Windows MobileTM, or the Symbian® operating system.
  • the operating system may include, or interface with a Java virtual machine module that enables control of hardware components and/or operating system operations via Java application programs.
  • Memory 304 further includes one or more data storage 310, which can be utilized by mobile device 300 to store, among other things, applications 320 and/or other data.
  • data storage 310 may also be employed to store information that describes various capabilities of mobile device 300. The information may then be provided to another device based on any of a variety of events, including being sent as part of a header during a communication, sent upon request, or the like.
  • Data storage 310 may also be employed to store social network" ' " ⁇ information including address books, buddy lists, aliases, user profile information, or the like.
  • Data storage 310 may further include program code, data, algorithms, and the like, for use by a processor, such as processor 302 to execute and perform actions.
  • data storage 310 might also be stored on another component of mobile device 300, including, but not limited to, non-transitory processor-readable removable storage device 336, processor-readable stationary storage device 334, or even external to the mobile device.
  • Data storage 310 may include, for example, raw data 312, and index data 314.
  • Applications 320 may include computer executable instructions which, when executed by mobile device 300, transmit, receive, and/or otherwise process instructions and data.
  • Applications 320 may include, for example, Cloud Search application 322, and Indexing
  • Application 324 Other examples of application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth.
  • VOIP Voice Over Internet Protocol
  • FIGURE 4 shows one embodiment of network device 400 that may be included in a system implementing the invention.
  • Network device 400 may include many more or less components than those shown in FIGURE 4. However, the components shown are sufficient to disclose an illustrative embodiment for practicing the present invention.
  • Network device 400 may represent, for example, one embodiment of at least one of network device 112, 114, or 120 of FIGURE 1.
  • network device 400 includes a processor 402 in communication with a memory 404 via a bus 428.
  • Network device 400 also includes a power supply 430, network interface 432, audio interface 456, display 450, keyboard 452, input/output interface
  • Power supply 430 provides power to network device 400.
  • Network interface 432 includes circuitry for coupling network device 400 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the Open Systems Interconnection model (OSI model), global system for mobile
  • GSM Global System for Mobile communications
  • CDMA code division multiple access
  • TDMA time division multiple access
  • UDP user datagram protocol
  • TCP/IP transmission control protocol/Internet protocol
  • SMS Short Message Service
  • MMS Multimedia Messaging Service
  • GPRS general packet radio service
  • WAP ultra wide band
  • WiMax IEEE 802.16 Worldwide Interoperability for Microwave Access
  • SIP/RTP Session Initiation Protocol/Real-time Transport Protocol
  • Network interface 432 is sometimes known as a transceiver, transceiving device, or network interface card (NIC).
  • Network device 400 may optionally communicate with a base station (not shown), or directly with another computing device.
  • Audio interface 456 is arranged to produce and receive audio signals such as the sound of a human voice.
  • audio interface 456 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others and/or generate an audio
  • a microphone in audio interface 456 can also be used for input to or control of network device 400, for example, using voice recognition.
  • Display 450 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computing device.
  • Display 450 may be a handheld projector or pico projector capable of projecting an image on a wall or other object.
  • Network device 400 also may also comprise input/output interface 438 for
  • Input/output interface 438 can utilize one or more wired or wireless communication technologies, such as USBTM, FirewireTM, WiFi, WiMax, ThunderboltTM, Infrared, BluetoothTM, ZigbeeTM, serial port, parallel port, and the like.
  • wired or wireless communication technologies such as USBTM, FirewireTM, WiFi, WiMax, ThunderboltTM, Infrared, BluetoothTM, ZigbeeTM, serial port, parallel port, and the like.
  • Human interface components can be physically separate from network device 400, allowing for remote input and/or output to network device 400.
  • information routed as described here through human interface components such as display 450 or keyboard 452 can instead be routed through the network interface 432 to appropriate human interface components located elsewhere on the network.
  • Human interface components include any component that allows the computer to take input from, or send output to, a human user of a computer.
  • Memory 404 may include Random Access Memory (RAM), Read-Only Memory (ROM), and/or other types of memory.
  • Memory 404 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Memory 404 stores a basic input/output system (BIOS) 408 for controlling low-level operation of network device 400.
  • BIOS basic input/output system
  • the memory also stores an operating system 406 for controlling the operation of network device 400. It will be appreciated that this component may include a general-purpose operating system such as a version of UNIX, or LINUXTM, or a specialized operating system such as Microsoft
  • Memory 404 further includes one or more data storage 410, which can be utilized by network device 400 to store, among other things, applications 420 and/or other data.
  • data storage 410 may also be employed to store information that describes various capabilities of network device 400. The information may then be provided to another device based on any of a variety of events, including being sent as part of a header during a
  • Data storage 410 may also be employed to store social networking information including address books, buddy lists, aliases, user profile information, or the like. Data stores 410 may further include program code, data, algorithms, and the like, for use by a processor, such as processor 402 to execute and perform actions. In one embodiment, at least some of data store 410 might also be stored on another component of network device 400, including, but not limited to, non-transitory media inside processor- readable removable storage device 436, processor-readable stationary storage device 434, or any other computer-readable storage device within network device 400, or even external to network device 400. Data storage 410 may include, for example, raw data 412, and indexed data 414.
  • Applications 420 may include computer executable instructions which, when executed by network device 400, transmit, receive, and/or otherwise process messages (e.g., SMS,
  • Multimedia Messaging Service MMS
  • IM Instant Message
  • email and/or other— ⁇ — ⁇ audio, video, and enable telecommunication with another user of another mobile device.
  • application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth.
  • Applications 420 may include, for example, cloud search application 422, and indexing application 424.
  • FIGURE 5 depicts an embodiment where the top layer of the figure, indicated as Web Application Layer 502, may represent a cloud-based front end that may abstract and enable access to the underlying cloud-hosted elements and objects that may be part of a multi-tenant application that may comprise Core User (e.g., Core User 518 and Core User 520), Roles (e.g., Role 522 and Role 524), and Apps (e.g., App 526 and App 528), indicated as the Search Head 504.
  • the objects in the Search Head 504 layer may be employed to access the Index objects located in the Indexers 506 layer of FIGURE 5.
  • the two projects, Project 1 (i.e., project 508) and Project 2 (i.e., project 510) in the Web Application Layer 502 may be examples of projects that have been elastically scaled to provide resources that may meet the specific needs of each project. As depicted, Project 1 may require twice as much performance and Project 2, the figure represents this with the additional Apps (i.e., App 526 and App 528) in the Search Head 504 layer, and the two index stores (i.e., index 530 and index 532) in the Indexer 506 layer.
  • FIGURE 5 demonstrates that a Frontend User may have access to one or more Projects.
  • frontend user 512 may have access to project 508, Frontend User 516 may have access to project 510, and Frontend User 514 may have access to both project 508 and project 510.
  • An embodiment may enable the Search Head 504 layer of a Project (e.g., project 508) to create unique Core Users in the Search Head 504 layer that may be linked to Frontend Users from the Web Application Layer 502.
  • Core User 518 may be linked to Frontend User 512 and Core User 520 may be linked to Frontend User 514.
  • a Role of the Core User that may be linked to the Frontend User may enable access control for the Apps and other objects within the Search Head layer.
  • Role 522 of Core User 518 may enable Frontend User 512 to access control for App 526 and App 528.
  • Role 524 of Core User 520 may enable Frontend User 514 to access control for App 526 and App 528.
  • the Role may determine the specific features that are made available to a Core User, and a Frontend User may have its access within a Project controlled by the Role of the Core User it is linked with. In other words, as depicted by FIGURE 5, the Roles may be thought of as providing pathways for the Frontend User to reach the Apps.
  • the Roles may also control access to indexers.
  • Role 522 may control access of Core User 518 to Apps 526 and/or 528 as well as index 530 and/or 532.
  • FIGURE 6 depicts an embodiment where Frontend User 1 (i.e., Frontend User 612) may be the owner of two projects, Project 1 (i.e., Project 608) and Project 2 (i.e., Project 610). As shown in the figure, Frontend User 2 (i.e., Frontend User 614) may not be an owner of a project. But, Frontend User 2 illustrates how a Frontend User may be invited by a project owner, represented in this embodiment by the depicted Frontend User 1. In at least one embodiment, Frontend User 612 and/or Frontend User 614 may be an embodiment of a Frontend User from FIGURE 5. In some embodiments, Project 608 and Project 610 may be embodiments of Project 508 and Project 510 of FIGURE 5, respectively. Web Application Layer 602, Search Head 604, and Indexers 606 may be embodiments of Web Application Layer 502, Search Head 504, and Indexers 506 of FIGURE 5, respectively.
  • Web Application Layer 602, Search Head 604, and Indexers 606 may be embodiments of Web Application Layer 502, Search Head
  • this embodiment may enable unique Core Users to be created and linked to Frontend Users.
  • Core User 618 and Core User 620 may be linked to Frontend User 612 and Core User 622 may be linked to Frontend User 614.
  • a Role may be created to define the features that may be accessed by the Frontend User.
  • Role 624 of Core User 618 may define the features that may be accessed by Frontend User 612 (e.g., App 626).
  • the embodiment depicted in FIGURE 6, may also enable Indexes (e.g., Index 632), Ports (e.g., Port 634), and associated storage to be located separately from the Search Head/Core layer objects. Also, embodiments may locate Indexes, and Ports across more than one computing device e.g., computing device 628 and computing device 630), including, network computing devices, blade servers, individual blade computers, cloud-based compute instances, virtual machines, and the like. Further, Indexes that may be associated with different Projects may be co-located on the same computing device, such as, network computing devices, blade servers, individual blade computers, cloud-based compute instances, virtual machines, and the like.
  • FIGURE 7 illustrates a logical flow diagram generally showing one
  • process 700 of FIGURE 7 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 700 or portions of process 700 of FIGURE 7 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 700 or portions of process 700 of FIGURE 7 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 700 begins, after a start block, at block 702 where at least one project may be initialized for at least one Frontend User.
  • at block 702 where at least one project may be initialized for at least one Frontend User.
  • Roles, Apps, Indexes, or the like may be created and/or configured for each unique project.
  • blocks 704, 706, 708, 710, and 712 may be performed in an order other than illustrated in FIGURE 7.
  • the associated Apps may be identified by the Frontend User.
  • Process 700 continues at block 706, where at least one Role may be determined for the project.
  • an operator and/or the Frontend User may generate and/or modify the access permissions associated with the Role.
  • at least one Role for each frontend user may be determined that may establish permissions for interacting with the at least one project and the at least one application for each Frontend User.
  • the Roles may configure access control for a Frontend User to the Apps and/or the indexers. For example, in some embodiments, Roles may enable one Frontend User to access one App (and/or indexer ), but not another App (and/or indexer).
  • a plurality of Roles may be determined for each Core User.
  • Process 700 proceeds to block 708, where a Core User on the project may be created for the Frontend User.
  • creating a Core User may include determining a Core User that corresponds to a frontend user and a project, wherein the Core User enables the Frontend User to interact with the project.
  • the Core User may inherit one or more corresponding Roles determined at block 706.
  • Process 700 continues next at block 710, where the Frontend User may be linked with the Core User and the Role.
  • an available index store may be established for the project.
  • an available index store may refer to an index store that is accessible by the Frontend User.
  • the Frontend User may be enabled to search the available index store.
  • the available index store may include one or more indexer instances.
  • the indexers may be configured based on the Roles. As described above, access to the index store (and/or the Apps) may be controlled by the Roles. In some other embodiments, a port may be assigned to each individual index.
  • the index store may be dynamically scaled, which is described in more detail below in conjunction with FIGURE 8. Briefly, however, the index store may be dynamically scaled based on the Frontend User account. For example, if the index store exceeds a pre -purchased data volume, then incoming data may continue to be index, but the Frontend User may not be able to access the incoming data until the Frontend User pays for an increase in the data volume.
  • the Frontend User may be enabled to select when and/or how dynamic scaling may be employed. For example, in some embodiments, the Frontend User may determine that if an index size limit is reached, at least one of the following may occur, including: automatically deleting additional data provided for the project beyond the size limit, the indexers may automatically stop indexing data, automatically increasing the storage size of data supported for a project, or the like.
  • process 700 may return control to a calling process.
  • FIGURE 8 illustrates a logical flow diagram generally showing one embodiment of a process for enabling access to a determined volume of data and restricting access to incoming data that exceeds the determined volume.
  • process 800 of FIGURE 8 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 800 or portions of process 800 of FIGURE 8 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 800 or portions of process 800 of FIGURE 8 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 800 begins, after a start block, at block 802, data may be provided.
  • the data may be incoming data for a given project.
  • the data may be archived data for a given project that may be recovered.
  • Process 800 continues at block 804, where the data may be indexed.
  • Process 800 proceeds to decision block 806, where a determination may be made whether a data volume limit is exceeded by the indexed data.
  • each project may include a data volume limit.
  • each frontend user (or group of frontend users) may be associated with a data volume limit.
  • the total stored data plus the indexed data may be compared to a maximum allowable data volume limit.
  • the data volume limit may be based on an amount of storage space purchased by a user (e.g., a Frontend User). If the data volume limit is exceeded, then process 800 may flow to block 810; otherwise, process 800 may flow to block 808.
  • the indexed data may be stored in an available index store.
  • an available index store may refer to an index store that is accessible to the user.
  • Process 800 may proceed to block 818.
  • process 800 may flow from decision block 806 to block 810.
  • the indexed data may be stored in an overflow index store.
  • the overflow index store may be unavailable to the user.
  • the overflow index store allows continued indexing of a user's data even if the amount of data has exceeded a purchased maximum data volume. However, the user may not have access to the data in the overflow index store, unless the user purchases additional data volume.
  • the overflow index store may be a FIFO buffer.
  • the data in the overflow index store may remain in the overflow index store until a user purchases additional data volume and/or the data is removed (e.g., deleted/archived) after a predetermined amount of time.
  • process 800 next flows to decision block 812, where a determination may be made whether access to data in the overflow index store is enabled. In some embodiments, access may be enabled if the user increases the data storage volume, such as by purchasing additional data storage. In at least one embodiment, the data storage volume may be elastically scaled as described in more detail below in conjunction with FIGURES 11-23. If access to the overflow index store is enabled, process 800 may flow to block 814; otherwise, process 800 may flow to block 818.
  • an amount of data in the overflow index store to make available may be determined.
  • the amount of data may be equal to or less than an increase in data storage volume. For example, if the user purchases an extra 2 GB of data storage, then it may be determined that up to 2 GB of data from the overflow index store may be made available.
  • the determined amount of data may be pulled from the overflow index store based on a FIFO (first in first out) buffer architecture of the overflow index store.
  • the indexed data that exceeds the data volume limit of the available index store is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit.
  • Process 800 proceeds to block 816, where the determined amount of data may be provided to the available index store.
  • providing the data from the overflow index store to the available index store may include modifying an index store flag; copying the data from the overflow index store to the available index store, or the like.
  • At least a portion of the indexed data in the overflow index store may be copied to archival storage.
  • the indexed data provided from the overflow index store to the available index store may also be provided to and/or copied to archival storage.
  • Process 800 continues next at block 818, where the user may be enabled to access the available index store. In at least one embodiment, the user may be enabled to perform searches for data in the available index store.
  • process 800 may return control to a calling process.
  • FIGURE 9 illustrates a logical flow diagram generally showing one embodiment of a process for managing an overflow index store.
  • process 900 of FIGURE 9 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 900 or portions of process 900 of FIGURE 9 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 900 or portions of process 900 of FIGURE 9 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 900 begins, after a start block, at decision block 902, where a determination may be made whether to remove data from the overflow index store.
  • data may be removed from the overflow index store when it is moved to the available index store.
  • data may be removed from the overflow index store after a predetermined amount of time.
  • data may be removed from the overflow index store if the amount of data in overflow index store exceeds a maximum overflow index store volume. If data is removed from the overflow index store, then process 900 may flow to decision block 904; otherwise, process 900 may return control to a calling process.
  • a determination may be made whether the removed data is moved to archival storage. In some embodiments, this determination may be based on user preferences, settings, rules (e.g., Roles), or the like. For example, an account for the user may indicate that the data removed from the overflow index store may be deleted without archiving the data. In other embodiments, the determination may be based on whether the data was added to the available index store at block 816 of FIGURE 8. In one such embodiment, if the data was added to the available index store, then the data may be moved to archival storage. If the data is moved to archival storage, then process 900 may flow to block 906; otherwise, process 900 may flow to block 908.
  • process 900 may flow to block 906; otherwise, process 900 may flow to block 908.
  • data may be moved from the overflow index store to archival storage.
  • data may be moved to the archival storage based on a FIFO buffer architecture of the overflow index store.
  • an amount of data to be moved to the archival storage may be based on the FIFO buffer.
  • the amount of data to be moved may be based on an amount of data added to the available index store at block 816 of FIGURE 8.
  • Process 900 proceeds to block 908, where data may be deleted from the overflow index store.
  • data may be deleted based on a FIFO buffer architecture of the overflow index store.
  • an amount of data to be deleted may be based on the FIFO buffer.
  • the amount of data to be deleted may be based on an amount of data added to the available index store at block 816 of FIGURE 8.
  • process 900 may return control to a calling process.
  • FIGURE 10 illustrates a logical flow diagram generally showing one embodiment of a process for enabling a project owner to invite another user into the project.
  • process 1000 of FIGURE 10 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1000 or portions of process 1000 of FIGURE 10 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
  • process 1000 or portions of process 1000 of FIGURE 10 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 1000 begins, after a start block, at block 1002, where a project owner may be enabled to send an invitation to an invited user. Also, any project user, or administrative user having the authority to invite users to a particular project may send invitations. Further, though not depicted in FIGURE 10, administrators, or other authorized users may define negative or positive restrictions on the users that may be invited to project. Restrictions may be defined for many purposes including, project security, privacy, licensing, and the like. Restrictions may be based on individual users as well as groups, lists, patterns, and the like.
  • Process 1000 proceeds to decision block 1004, where a determination may be made whether a user accepted the invitation.
  • an invited user may have the choice to accept the invitation. However, some embodiments may automatically join invited users to a project without giving them the option to decline the invitation. If an invited user declines or ignores an invitation, then process 1000 may return to the calling process; otherwise, process 1000 may flow to decision block 1006. In some embodiments the sender of the invitation may receive notice that the invitation has been declined or ignored.
  • the invited user may create a Frontend User account.
  • a new Core User on a project may be creased. Proceeding to block 1012, a corresponding new Role for the new Core User may also be determined. In some embodiments, one or more new Roles may be created for the new Core User. In other embodiments, the new Core User may inherit one or more Roles from the user who initiated in invitation.
  • Process 1000 proceeds to block 1014, where the invited Frontend User may be linked with the new Core User and the new Role. After block 1014, process 1000 may return control to a calling process.
  • FIGURE 11 illustrates a logical flow diagram generally showing one embodiment of an overview process for elastically scaling a project.
  • process 1100 of FIGURE 11 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1100 or portions of process 1100 of FIGURE 11 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1100 or portions of process 1100 of FIGURE 11 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 1100 begins, after a start block, at block 1102, where resources allocated for use by the Project may be determined to be sub-optimal. Such a determination may be made by many methods, including, rule -based systems, or manual systems. Embodiments may apply various rule-based systems that may enable determining that a project is operating with suboptimal resources. Such rule-based systems may include, detecting if a performance metric threshold has been exceeded, triggering resource allocations based on schedule, triggering resource allocations based on pattern prediction that may indicate a potential change in utilization requirements, and the like. In addition, an operator may determine by observation that resources are suboptimal.
  • optimal level of project resources may be determined.
  • optimal resources may be static or dynamic.
  • optimal resources may be automatically determined by rule-based systems, predictive computing techniques, or the like.
  • optimal resources may be manually determined by an operator. The specific allocation of resources may be determined in part by the characteristics of the project, or they may be determined by factors external to project. For example, a project that requires a very large dataset may have an optimal resource allocation that emphasizes increased data storage volume over processor performance.
  • Process 1100 continues at block 1106, where the project resources may be scaled to support the determined optimal level. Additional non-limiting embodiments of elastic scaling of project resources are described below. After block 1106, process 1100 may return control to a calling process.
  • FIGURE 12 illustrates a logical flow diagram generally showing one embodiment of a process for receiving resource allocation direction from an operator manipulating a user interface control.
  • process 1200 of FIGURE 12 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1200 or portions of process 1200 of FIGURE 12 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1200 or portions of process 1200 of FIGURE 12 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • the user interface control may have a user interface that represents a slider, dial, selection lists, radio buttons, check boxes, and the like.
  • Embodiments may be arranged to map various elastic scaling dimensions to a control.
  • a control may represent processing performance, index storage, time-window of received data, and the like.
  • An embodiment of a use case of a slider user interface is described in more detail below in conjunction with FIGURE 23.
  • Process 1200 begins, after a start block, at block 1202, where a control change (delta) may be received.
  • a control change may be received.
  • an operator may change, or manipulate the control to provide the delta.
  • a signal, message, or notification from the control may be received indicating that a change in the control may have occurred (i.e., the delta).
  • Process 1200 proceeds next to decision block 1204, where a determination may be made whether the delta is greater than zero.
  • a delta greater than zero may indicate a request to increase resources for the instant project. If the delta is greater than zero, then process 1200 may flow to block 1206; otherwise, process 1200 may flow to decision block 1208.
  • the resources for the project may be increased.
  • One embodiment of a process for increasing resources for a project is described in more detail below in conjunction with FIGURE 13. After block 1206, process 1200 may return control to a calling process.
  • processing may flow from decision block 1204 to decision block 1208.
  • decision block 1208 a determination may be made whether the delta is less than zero. In at least one embodiment, a delta less than zero may indicate a request to decrease resources for the instant project. If the delta is less than zero, then process 1200 may flow to block 1210; otherwise, process 1200 may return control to a calling process.
  • the resources for the project may be decreased.
  • One embodiment of a process for decreasing resources for a project is described in more detail below in conjunction with FIGURE 10.
  • the concept of increasing or decreasing resources as used herein is generalized to refer to increasing or decreasing project resources in a defined and configured "dimension."
  • increasing a slider (delta > zero) means to emphasize storage and de-emphasize processing speed, then the effect may be to increase storage and reduce the CPU quota.
  • decreasing the slider (delta ⁇ zero) may mean to de-emphasize storage and emphasize processing speed, then the effect may be to de-increase storage and increase the CPU quota.
  • sliders and controls may be arranged and configured in many ways to provide operators with a user-interface for controlling the elastic scaling of project resources.
  • Embodiments may range from single controls that effect many elements of the computing process with a single action, to many fine grain controls that may be designed to elastically scale individual computing performance attributes.
  • an embodiment enabling fine grain control may have separate controls to adjust CPU quota, RAM use, disk access, network bandwidth, process priority, and the like.
  • an embodiment may implement a single high level control in the aggregate that may be arranged to hide and abstract the separate adjustments to CPU quota, RAM use, disk access, network bandwidth, process priority, and the like.
  • the operator may not be required to have knowledge that a single change in the high level control may affect many of the project's resource components.
  • process 1200 may return control to a calling process.
  • FIGURE 13 illustrates a logical flow diagram generally showing one embodiment of a process for increasing a project's resources as part of elastic scaling.
  • process 1300 of FIGURE 13 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1300 or portions of process 1300 of FIGURE 13 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1300 or portions of process 1300 of FIGURE 13 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 1300 begins, after a start block, at block 1302, where particular computing resources that correspond to the requested increase in resources may be determined.
  • a requested increase of resources may be a simple request to increase a single component of the computing system, such as CPU quota, or it may represent an increase in an elastic scaling dimension that may require the adjustment and reallocation of more than one component in the computing system, such as, CPU quota, RAM quota, network bandwidth quota, and the like.
  • an increase of one resource may result in a decrease of another resource.
  • Process 1300 next proceeds to block 1304, where the impact that the requested resources may have on the project, other projects, and/or the system in general may be determined.
  • the impact of the requested resources may be static or dynamic.
  • the impact may be automatically determined by rule-based systems, predictive computing techniques, or the like.
  • the impact may be manually determined by an operator.
  • Embodiments may be arranged to define rules and thresholds that may set limits on the resources that may be allocated to individual projects. Also, rules may be arranged to determine minimum resources that may be made available for other projects in the system.
  • a request to increase index storage size to a very large size may cause such a large data set to be created that the project may not have CPU quota, or RAM quota to effectively process the large dataset. How a condition like this is handled depends on how the embodiment is arranged and configured. In some cases, such an increase in project resources may be aborted because it is out of bounds. In other cases, referring to the example above, an embodiment may determine the necessary CPU quota and RAM quota that may be necessary to keep up with the projected increased index storage size.
  • process 1300 continues at decision block 1306, where a determination may be made whether there is global headspace available for the requested increase of resources. In at least one embodiment, this determination may be made by testing the amount of global headspace available. This test may take into account all of the resources available and provisi— J r ⁇ iL ⁇ operator or process that is requesting the increase of resources for the project. For example, an embodiment may require users to supply a pre-authorization when requested resources would increase a user's total usage of system resources beyond a certain level. Or, an embodiment may be configured to require an administrative user to authorize any resource increase request that may increase total resource use beyond a configured level. In any event, an embodiment may test whether the global conditions may be able to absorb the requested increase in resources.
  • the requested increase in the resource may not be executed at the time of the request.
  • process 1300 may flow to decision block 1308; otherwise, process 1300 may flow to block 1312.
  • the increase in resources would exceed global thresholds and/or limitations, then the request for the increase of resources may be discarded.
  • Process 1300 continues at decision block 1308, where a determination may be made whether the local resources related to the project may be capable of handling the requested increase in project resources. For example, in an embodiment that may be employing virtual machines, the requested increase in project resources may require that a virtual machine allocate more virtual disk space, or it may require that the virtual machine be migrated to another physical machine that may be able to accommodate the requested increase in project resources. In some embodiments, if such operations are unable to be resolved immediately (e.g., within a predefined time period), then the local resources may not be capable of handling the requested increase of project resources. If the local resources are capable of handling the requested increase in project resources, then process 1300 may flow to decision block 1310; otherwise, process 1300 may flow to block 1312 to queue the request.
  • the local resources related to the project may be capable of handling the requested increase in project resources. For example, in an embodiment that may be employing virtual machines, the requested increase in project resources may require that a virtual machine allocate more virtual disk space, or it may require that the virtual machine be migrated to
  • a determination may be made whether the requested increase in project resources may be arranged or configured to occur at a scheduled time in the future.
  • a user and/or operator may determine a scheduled time when the requested increase may occur.
  • the scheduled time may be predetermined, such as five minutes from when the request was received.
  • the scheduled time may be based on the resources requested.
  • process 1300 may flow to block 1312 to queue the request; otherwise, process 1300 may flow to block 1314.
  • the request for the increase of resources may be submitted to a queue.
  • the request may remain in the queue until conditions allow it to be fulfilled (i.e., available resources).
  • conditions allow it to be fulfilled i.e., available resources.
  • queuing project resource requests is described in more detail below in conjunction with FIGURE 15.
  • a process or operator requesting an increase in project resources may not have knowledge of the global resource allocation. Queuing the requested increase in project resources may enable short running processes to finish and expire so that their released resources may be made available to a queued request for increasing project resources.
  • process 1300 may return control to a calling process.
  • process 1300 may flow to block 1314.
  • the determined resources for the requested increase may be allocated. Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
  • process 1300 may return control to a calling process.
  • FIGURE 14 depicts a logical flow diagram generally showing one embodiment of process for responding to a requested decrease in project resources.
  • process 1400 of FIGURE 14 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1400 or portions of process 1400 of FIGURE 14 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1400 or portions of process 1400 of FIGURE 14 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • a requested decrease in project resources may require some system or computing resources to decrease while others may be required to be increased depending on how the relevant control dimension has been arranged and configured.
  • Process 1400 begins, after a start block, at decision block 1402, where a determination may be made whether the requested decrease in project resources may be scheduled to occur in the future.
  • decision block 1402 may employ embodiments of decision block 1310 of FIGURE 13 to determine if a request may be scheduled to occur at a time after the request is received. If the decrease in project resources may be scheduled in the future, then process 1400 may flow to block 1410 to queue the request at least until the scheduled time; otherwise, process 1400 may flow to block 1404.
  • block 1404 may employ
  • block 1304 of FIGURE 13 determines an impact of resources on the project and the system (noting that block 1304 of FIGURE 13 determines based on an increase in resources and block 1404 determines based on a decrease in resources).
  • Process 1400 next proceeds to decision block 1406, where a determination may be made whether the impact of the requested decrease in project resources is unacceptable.
  • running processes of the current project may be analyzed to determine if they can run to completion under the conditions of the decreased resources.
  • a current project may not run to completion if the decrease in resources starves the project of processor time, which may be determined to be unacceptable.
  • process 1400 may flow to block 1408; otherwise, process 1400 may flow to block 1410 to queue the request, which may enable the current processes to finish gracefully.
  • a requested decrease in project resource may be arranged and configured so it may not cooperate with the running processes. In such cases, an embodiment may terminate the running process to avoid unpredictable outcomes that may occur if the processes try to continue to run reduced resource environment.
  • the resource decrease request may be queued.
  • An embodiment of a project resource request queue is shown in FIGURE 15. As discussed above project resource requests may be submitted to a request queue if the request may not be completed at the time it may be initiated. This may occur for many reasons, including lack of local or global resources, lack of authorizations that may be needed, the request may be deliberately schedule to occur in the future, currently running processes may be blocking the request, and the like.
  • a project resource request queue may be arranged to hold request records that may comprise a data structure that carries the metadata that may contain the information about the resources that may have been determined necessary to execute the request.
  • metadata may include instructions on how to allocate resources for all of the components impacted by the resource request.
  • a project resource request may require a CPU quota adjustment and index storage size allocation as part of a single project resource request.
  • a queued resource request may contain metadata and instructions that may be used to executed the CPU quota change and the index storage allocation.
  • the queued request may contain the control delta information from the component that initiated the project request and the underlying components and instructions needed to execute the request may be determined at the time of allocation.
  • process 1400 may flow from decision block 1406 to block 1408.
  • resource allocation for the current project may be decreased. Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
  • FIGURE 15 illustrates a logical flow diagram generally showing one embodiment of process for queuing resource requests.
  • process 1500 of FIGURE 15 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1500 or portions of process 1500 of FIGURE 15 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1500 or portions of process 1500 of FIGURE 15 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 1500 begins, after a start block, at decision block 1502, where a determination may be made whether there are queued requests. In at least one embodiment, this determination may be based on whether the queue is holding any requests. If the queue is holding requests, then process 1500 may flow to block 1504; otherwise, process 1500 may return control to a calling process.
  • a next request in the queue may be examined.
  • this request examination may include a determination of the resources associated with the next request and a determination of the conditions that need to be met for the next request to be acted upon (e.g., acceptable resources available to accommodate the request).
  • the conditions may include the conditions that led to the next request being entered into the request queue in the first place.
  • the condition may include a scheduled time.
  • Process 1500 next proceeds to decision block 1506, where a determination may be made whether the conditions to execute the next request are satisfied for the next request. For example, a next request that required prior authorization before proceeding may be tested to see if sufficient authorization has been obtained.
  • global resource conditions may be examined to determine if a next request that may have been queued because of a lack of available global resources now may have sufficient resources available to proceed. For next requests that may have been queued so they may occur at a scheduled time in the future the time may be checked. If the conditions for the next request are met, then process 1500 may flow to block 1508;
  • process 1500 may loop to decision block 1502. If conditions for the next request are not met, the process may leave the request in the queue and may examine the next request in the queue.
  • a queue may be configured such that if a request remains in the queue longer that a specified time an additional action may be taken. Such actions may include, notifying the operator, generating alerts in the form of log messages or other notifications, removing the request from the queue, and the like.
  • the determined resources based on the next request may be allocated.
  • Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
  • Process 1500 continues at block 1510, where the completed request may be removed from the queue.
  • the next request may be configured to be returned to the queue after it has executed. Some requests may be scheduled to occur at a certain time each day. These requests may be returned to the queue to wait until the next time to execute. Other requests may be considered "standing" requests that may remain in the queue waiting for certain conditions to be met.
  • an embodiment may be arranged to automatically increase index storage as it is needed. In this kind of case, an embodiment may use a queued request that may execute when it determines that more index storage is needed.
  • Embodiments may be arranged to have multiple project resource request queues designated for different types of requests. Also, embodiments may employ general purpose queues that may be used for queue project resource requests as well as other types of requests and events. After block 1510, process 1500 may return control to a calling process.
  • FIGURE 16 depicts a logical flow diagram generally showing one embodiment of process for elastically scaling based on index storage size.
  • process 1600 of FIGURE 16 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1600 or portions of process 1600 of FIGURE 16 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
  • process 1600 or portions of process 1600 of FIGURE 16 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B
  • Process 1600 begins, after a start block, at block 1602, where an optimal storage size for a new index store may be determined.
  • optimal storage size may be static or dynamic.
  • optimal storage size may be automatically determined by rule-based systems, predictive computing techniques, or the like.
  • optimal storage size may be manually determined by an operator. The optimal size may be determined by a rule-based system tailored to the specific project that owns the index store.
  • An embodiment of a rule-based system for elastic scaling may employ triggers linked to performance and resource thresholds.
  • the optimal size for index store may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator has used a control to determine the optimal size, an embodiment may translate the operation of the control into the actual optimal index store size. For example, an operator may use a user interface to control to increase an index store from 5 gigabytes to 10 gigabytes. In this case an embodiment may have to make several determinations to account for the actual storage size needed to account for administration overhead that may not be apparent to the operator.
  • Process 1600 continues at block 1604, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the index store. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
  • Process 1600 proceeds next to block 1606, where new objects and/or processes may be allocated and/or instantiated to optimally service the scaled new index store.
  • Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
  • the new index store may be determined. At least one embodiment for determining a new index store is described in more detail below in conjunction with FIGURE 19. Briefly, however, in some embodiments indexers may be activated (i.e., marked as read/write) or deactivated (i.e., marked as read-only based on a number of indexers needed for the scaled new index store storage. In some other embodiments, the current index store may be marked as read-only and the new index store may be generated from new indexers. This may ensure that the all new writes to the index store may be written into the scaled new index store storage.
  • Process 1600 next proceeds to block 1610, which is discussed in more detail below in conjunction with FIGURE 22.
  • an input proxy may be directed to use the scaled new index store for the indexing of incoming events.
  • Process 1600 continues at block 1612, where the new objects and/or processes may be employed to index incoming events using the scaled new index store.
  • the new objects and/or processes may be employed to index incoming events using the scaled new index store.
  • incoming events and/or incoming data may include new events/data and/or archived events/data.
  • the scaled new index store may be utilized based on Roles of the new indexes.
  • new indexes (which may include the new objects, new processes, or the like) may be automatically added to existing indexes (which may include the existing objects, existing processes, or the like).
  • the new indexes may be attached to and/or inherit Roles associated with the existing indexes for at least one Core User of the project.
  • the input proxy may utilize the inherited Roles during a scaling event to determine when to send incoming events (and/or other data) to the new indexes. As described above, the Roles may determine access controls to the indexes.
  • process 1600 may return control to a calling process.
  • FIGURE 17 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on indexing performance.
  • process 1700 of FIGURE 17 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1700 or portions of process 1700 of FIGURE 17 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1700 or portions of process 1700 of FIGURE 17 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Indexing performance may be a compound elastic scaling dimension that may be generally related to speed of the processing for a project.
  • multiple component resources such as CPU quota, network bandwidth quota, RAM quota, index storage, and like, may be scaled to execute elastic scaling based on Indexing Performance.
  • Process 1700 begins, after a start block, at block 1702, where resources needed for optimal indexing performance may be determined.
  • optimal indexing performance may be automatically determined by rule-based systems, predictive computing techniques, or the like.
  • optimal indexing performance may be manually determined by an operator.
  • the specific resources needed may be determined by a rule-based system tailored to a specific project.
  • An embodiment of a rule -based system for elastic scaling may employ triggers linked to performance and resource thresholds. Different projects may be configured to have different triggers and different thresholds depending on the requirements of the project.
  • the optimal resource allocation may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator may use a control to determine the optimal indexing performance an embodiment may translate the operation of the control into the actual optimal allocation of resources.
  • an operator may use a user interface to control to increase a projects indexing performance from “low” to "high.”
  • an embodiment may have to make several determinations to account for CPU quota, disk access quota, RAM quota, storage size, and the like, the details of which may not be apparent to the operator.
  • Process 1700 proceeds to block 1704, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the indexing performance.
  • An embodiment may allocate and format a new storage even if the size of the data store has not increase or decrease. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
  • Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
  • Process 1700 proceeds next at block 1708, where the new index store may be
  • block 1708 may employ embodiments of block 1608 of FGURE 16 to determine the new index store.
  • an input proxy may be directed to use the scaled new index store for the indexing of incoming events.
  • Process 1700 proceeds to block 1712, where the new objects and/or processes may be employed to index incoming events using the scaled new index store. In at least one
  • incoming events and/or incoming data may include new events/data and/or archived events/data.
  • block 1712 may employ embodiments of block 1612 of FIGURE 16 to use the scaled new index store.
  • process 1700 may return control to a calling process.
  • FIGURE 18 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on an event time window.
  • process 1800 of FIGURE 18 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 1800 or portions of process 1800 of FIGURE 18 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 1800 or portions of process 1800 of FIGURE 18 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Event time window may be a compound elastic scaling dimension that may be generally related to the index storage sizes for a project.
  • An embodiment may be arranged to receive data such that the index store may be sufficient to hold data received during a determined time window.
  • a time window may be described as date-time expression such as, a start date -time and an end date-time, a duration, such as 30 days, start date-time plus a duration, and the like.
  • multiple component resources such CPU quota, network bandwidth quota, RAM quota, index storage, and like, may be scaled to execute elastic scaling based on event time windows.
  • Process 1800 begins, after a start block, at block 1802, where a new event time window may be determined.
  • the event time window may be submitted as a complete time duration expression or it may be dynamically determined based on past event traffic and predicted future event traffic or any method to describe a time window for capturing events.
  • Process 1800 proceeds next to block 1804, where resources needed for indexing data for the determined event time window may be determined.
  • the specific resources needed may be determined by a rule-based system tailored for specific projects.
  • An embodiment of a rule-based system for elastic scaling may employ triggers linked to performance and resource thresholds. Different projects may be configured to have different triggers and different thresholds depending on the requirements of the project.
  • the optimal resource allocation may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator may use a control to determine the optimal indexing performance an embodiment may translate the operation of the control into the actual optimal allocation of resources.
  • process 1800 continues at block 1806, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the determined event time window in the new index store. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
  • Process 1800 proceeds to block 1808, where new objects and/or processes may be allocated and/or instantiated to optimally support the determined event time window.
  • Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
  • the new index store may be determined.
  • block 1708 may employ embodiments of block 1608 of FGURE 16 to determine the new index store.
  • an input proxy may be directed to use the scaled new index store for the indexing of incoming data.
  • Process 1800 continues to block 1814, where the new objects and/or processes may be employed for indexing incoming data using the scaled new index store.
  • incoming events and/or incoming data may include new events/data and/or archived events/data.
  • block 1814 may employ embodiments of block 1612 of FIGURE 16 to use the scaled new index store.
  • process 1800 may return control to a calling process.
  • FIGURE 19 illustrates a logical flow diagram generally showing one embodiment of a process for changing a number of indexers associated with an index store.
  • process 1900 of FIGURE 19 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1900 or portions of process 1900 of FIGURE 19 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
  • process 1900 or portions of process 1900 of FIGURE 19 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • changing a number of indexers may also include changing an amount of resources associated with a project, such as, for example, storage, objects, processors, or the like.
  • adding additional indexers may include adding additional storage space and/or processors (and/or other resources).
  • reducing a number of indexers may include reducing storage space and/or processors (and/or other resources.
  • resources for a given indexer server may also be changed.
  • resources of an indexer server may be changed by changing a number of indexes used by a project, which may also change an amount of storage, processing capacity, or the like for the project.
  • Process 1900 begins, after a start block, at decision block 1902, where a determination may be made whether additional indexers may be added to the index store. In at least one embodiment, the determination may be based on the elastic scaling. In some embodiments, if the storage size and/or the performance are increased, then indexers may be added to the index store. If indexers may be added to the index store, then process 1900 may flow to decision block 1904; otherwise, process 1900 may flow to decision block 1906.
  • a determination may be made whether to remove indexers from the index store. In at least one embodiment, the determination may be based on the elastic scaling. In some embodiments, if the storage size and/or the performance are decreased, then indexers may be removed from the index store. If indexers may be removed from the index store, then process 1900 may flow to block 1908; otherwise, process 1900 may return control to a calling process.
  • At block 1908, at least one indexer may be deactivated.
  • deactivating an indexer may include marking the indexer as read-only.
  • a read-only indexer may indicate that the indexer is not currently active for the new index store.
  • a determination may be made on which indexers to deactivate based on time (e.g., creation time, last modified time, last access time, or the like), size of the indexer, or the like.
  • a user may be prevented from accessing a deactivated indexer.
  • the indexer if the indexer is deactivated for a period of time, then the data may be removed from the deactivated indexer (e.g., deleted and/or moved to archival storage). If, at decision block 1902, additional indexers may be added to the index store, then processing may flow from decision block 1902 to decision block 1904. At decision block 1904, a determination may be made whether there are deactivated indexers (such as were deactivated at block 1908). If there are deactivated indexers, then process 1900 may flow to block 1912; otherwise, process 1900 may flow to block 1910.
  • At block 1912 at least one currently deactivated indexer may be activated and added to the new index store.
  • activating may include marking a read-only indexer as a read/write indexer.
  • process 1900 may return control to a calling process. If, at decision block 1904, there are no deactivated indexers, then process 1900 may flow from decision block 1904 to block 1910.
  • new indexers may be added to the new index store. After block 1910, process 1900 may return control to a calling process.
  • FIGURE 20 shows a logical flow diagram generally showing one embodiment of a process to recall archived data for re -indexing.
  • process 2000 of FIGURE 20 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 2000 or portions of process 2000 of FIGURE 20 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 2000 or portions of process 2000 of FIGURE 20 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Embodiments may be arranged to archive input data as it is indexed. (See, input proxy, as shown in FIGURE 22, for more details about archiving.) Data stored in the archives may be in the same format and form as it was sent to the input proxy.
  • a process may process data from the archive and create indexes from the data. This may enable operators to re-create indexes from the archived data and it may enable operators to create new indexes that may be processed differently from the indexes that may have been created when the data was first received.
  • Process 2000 begins, after a start block, at block 2002, where a time window of events may be determined for re-indexing.
  • the time windows may be determined be a rule-based algorithm that may automatically determine the time window or an operator may select the time window manually using a user interface.
  • Process 2000 continues at block 2004, where event data records may be retrieved from the archive.
  • Process 2000 proceeds next to block 2000, where the retrieved event data records are indexed using the current project's parameters. Recall of archived event data may occur in the context of a new project or it may occur as part of an existing project.
  • process 2000 may loop to block 2004; otherwise, process 2000 may return control to a calling process.
  • a new scaled index store may be employed based on elastic scaling.
  • FIGURE 21 illustrates a logical flow diagram generally showing one embodiment of a process for transitioning to scaled index stores.
  • process 2100 of FIGURE 21 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 2100 or portions of process 2100 of FIGURE 21 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 2100 or portions of process 2100 of FIGURE 21 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 2100 begins, after a start block, at block 2102, where event data may be received from an input proxy. Embodiments of employing an input proxy are described in more detail below in conjunction with FIGURE 22. Process 2100 next proceeds to block 2104, where the received event data may be indexed. In at least one embodiment, the indexed event data may be stored in the new scaled index store.
  • Process 2100 continues at block 2106, where the total storage for the new events stored in the scaled new index store and the old event that may be stored in the previously used readonly index store may be determined. Proceeding to decision block 2108, a determination may be made whether the total amount of storage used is greater than the amount of stated provisioned for the project. If the total storage used is greater than the amount of storage provisioned for the project, then process 2100 may flow to block 2110; otherwise, process 2100 may loop back to block 2102 to receive additional event data from the Input Proxy.
  • the index records in the previously provisioned storage may be marked for archival or removal.
  • process 2100 may flow to block 2112; for the index records marked for removal, process 2100 may flow to block 2114.
  • the index records marked for archival may be copied from the previously provisioned storage to archival storage.
  • index records marked for removal may be deleted from the previously provisioned storage.
  • index records marked from archive may be deleted from the previously provisioned storage after being copied to the archival storage.
  • process 2100 may return control to a calling process.
  • FIGURE 22 illustrates a logical flow diagram generally showing one embodiment of an input proxy operation.
  • process 2200 of FIGURE 22 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4.
  • process 2200 or portions of process 2200 of FIGURE 22 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4.
  • process 2200 or portions of process 2200 of FIGURE 22 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
  • Process 2200 begins, after a start block, at block 2202 where the proxy may receive data.
  • the data may be received from remote clients and sources.
  • the data source may stream the data to the input proxy using various networking technologies, including, Ethernet, 802.11 (a,b,g, and n), Bluetooth, TCP/IP, UDP, HTTP, internal specific point-to-point protocol, and the like.
  • the proxy may throttle (increase or decrease) a rate at which data incoming data is streamed.
  • the data rate may be based on a corresponding project. For example, the data rate may be based on an amount of storage space allocated for the project (e.g., an amount of storage space purchased by a user for the project).
  • the data may be received and a process may "tee" the data, sending it down at least two paths at the same time, which may be depicted in FIGURE 22 as Path A and Path B.
  • process 2200 may proceed from block 2202 to block 2204.
  • the raw data may be load balanced to determine which indexer to use.
  • Process 2200 continues next at block 2206, where the received data may be indexed on the load balanced indexers.
  • a size of each indexer may not be limited. In other embodiments, the size of each indexer may be limited. In one such embodiment, an amount of data per indexer may be monitored. In one embodiment, the monitoring may be asynchronous to incoming data. In some embodiment, if a limit for an indexer is reached or almost reached (e.g., less than 5% remaining), then a message may be sent to a user to indicate that the indexer limit is reached or almost reached. In some other embodiments, indexing may be disabled if a user exceeds purchased resource allocation.
  • indexing may be disabled if a user and/or project exceeds an allocated storage size (e.g., previously purchased amount of storage). In another embodiment, indexing may be disabled if a user discontinues paying for the indexing. In yet another embodiment, indexing may be disabled if a user and/or project exceeds a number of buckets for the project.
  • an allocated storage size e.g., previously purchased amount of storage.
  • Process 2200 next proceeds to block 2208, where the indexed data and event records may be stored in the project's index store.
  • process 2200 may proceed from block 2202 to block 2210.
  • the raw data may be sent to be copied to intermediate storage.
  • the intermediate storage may be arranged to be available, fast and reliable.
  • Embodiments may employ a FIFO buffer implemented by local hard disks, RAID arrays, and the like, to ensure that the raw data is not lost.
  • Process 2200 continues at block 2212, where the data may be archived as archival storage becomes available. Moving the raw data from the intermediate storage to the long term archive may be executed by a separate process that reliably moves the raw data from the intermediate storage to the long term storage. After block 2208 and 2212, process 2200 may return control to a calling process.
  • program instructions may be provided to a processor to produce a machine, such that the instructions, which execute on the processor, create means for implementing the actions specified in the flowchart block or blocks.
  • the computer program instructions may be executed by a processor to cause a series of operational steps to be performed by the processor to produce a computer-implemented process such that the instructions, which execute on the processor to provide steps for implementing the actions specified in the flowchart block or blocks.
  • the computer program instructions may also cause at least some of the operational steps shown in the blocks of the flowchart to be performed in parallel. Moreover, some of the steps may also be performed across more than one processor, such as might arise in a multi-processor computer system. In addition, one or more blocks or combinations of blocks in the flowchart illustration may also be performed concurrently with other blocks or combinations of blocks, or even in a different sequence than illustrated.
  • blocks of the flowchart illustration support combinations of means for performing the specified actions, combinations of steps for performing the specified actions and program instruction means for performing the specified actions. It will also be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by special purpose hardware-based systems, which perform the specified actions or steps, or combinations of special purpose hardware and computer instructions.
  • FIGURE 23 illustrates a non-exhaustive example of a use case of an embodiment of user-interface controls that may be employed to enable elastic scaling.
  • an embodiment may be arranged to provide user-interface controls, such as sliders, dials, radio buttons, spinners, touch controls, gesture controls, and the like, that may enable an operator to elastically scale resources that may be associated with a project.
  • user-interface controls such as sliders, dials, radio buttons, spinners, touch controls, gesture controls, and the like.
  • One exemplary user-interface is shown in the FIGURE 23. This may represent a user-interface having three slider controls for controlling the resource for the project.
  • a slider control (e.g., control 2302, control 2304, and/or control 2306) may generally comprise a slider 2308 and a slider thumb 2310.
  • An operator may manipulate the slider thumb 2310 by moving it forward or backwards lengthwise within the slider 2308.
  • the movement of the slider thumb 2310 may generate intermediate values that may be used for determining corresponding to elastic scaling values.
  • Control 2302 on the top of the figure, may be use for the elastic scaling of index storage size.
  • Control in the center of the figure, may be used for the elastic scaling of the number of server instances the may be used to support a project.
  • control 2306 at the bottom of the figure, may be used to set the index storage size by event time window.
  • Control 2302 may also include text box 2312 and text box 2314.
  • text box 2312 may display the current value represented by the position of slider thumb 2310 in slider 2308.
  • text box 2312 may enable a user to enter a desired value, and slider thumb 2310 may automatically slide to the corresponding position in slider 2308.
  • text box 2314 may display how much the current value (i.e., what is displayed in text box 2312) of control 2302 will cost the owner of the project. In some other embodiments, text box 2314 may enable a user to enter a desired cost, and slider thumb
  • each slider control may include text box 2312 and/or text box 2314.
  • embodiments are not restricted to arranging text boxes as depicted herein. Embodiments may be expected to use a variety of user interface methods to indicate the current value of the controls and well as how much the selected value may cost. Also, embodiments may be arranged to display values different and in addition to those depicted herein, or none at all.
  • Some embodiments may arrange for the user interface controls to use different scales for the variables different that may be represented by the control, such as, linear (as illustrated by control 2302), logarithmic (as illustrated by control 2306), parametric (as illustrated by control 2304), and the like.
  • An embodiment may be arranged to use a particular scale based on the type of user-interface being used and the resource that may be represented.
  • the scale of the control may be determined based on how the user may be charged for the underlying resource. For example, if a linear increase in the number of server instances many have a logarithmic increase in the cost to the user it may be advantageous to scale the control to account for the change in cost rather than the change in resources.
  • multiple controls may be linked together such that adjusting one control may cause another control to adjust as well.
  • embodiments may provide additional user controls (not shown), such as, checkboxes, that may enable a user to lock a control so it may remain in a fixed position as other controls are adjusted.
  • some controls may represent high level, generalized, or aggregated representations of resources, in this case a user may select a higher level control and "drill-down" into the underlying elements that the high level, generalized, or aggregated control may be representing.

Abstract

Embodiments are directed towards a system and method for a cloud-based front end that may abstract and enable access to the underlying cloud-hosted elements and objects that may be part of a multi-tenant application, such as a search application. Search objects may be employed to access indexed objects. An amount of indexed data accessible to a user may be based on an index storage limit selected by the user, such that data that exceeds the index storage limit may continue to be indexed. Also, one or more projects can be elastically scaled for a user to provide resources that may meet the specific needs of each project.

Description

DATA VOLUME MANAGEMENT
CROSS REFERENCE TO RELATED APPLICATIONS
This application claims priority to U.S. Patent Application, Serial No. 13/572,434 filed on August 10, 2012, entitled, "Data Volume Management," which claims the benefit of U.S. Provisional Patent Application, Serial No. 61/523,063 filed on August 12, 2011, entitled "Data Volume Management," both of which are incorporated herein by reference.
TECHNICAL FIELD
The invention is directed to providing services for use by a customer, and more particularly, to providing search services to the customer over a network with a cloud-based system.
BACKGROUND
Many companies install and maintain software for use on a variety of distributed systems that can range from a single computer for a small business to a collection of servers and a plurality of user computer nodes for a large corporation. Also, the volume of data processed by a node in a small business can be a magnitude less than the volume of data processed by a node in a large corporation. Further, at different times the amount of data to be indexed for searching and the need for speed to search the indexed data can vary.
Additionally, the type of data processed by large and small business can vary widely, such as sending and receiving messages, creating and storing documents, hosting web sites, database searches, facilitating online transactions, and the like. Furthermore, the expense of developing and maintaining software that can handle the type and volume of data processed by a large corporation can be substantially greater than the effort expended to do somewhat the same for a small business. Consequently, cloud-based systems that provide software as a service are increasingly popular with different sizes of businesses having different volumes of data and types of data. BRIEF DESCRIPTION OF THE DRAWINGS
Non-limiting and non-exhaustive embodiments are described with reference to the following drawings. In the drawings, like reference numerals refer to like parts throughout the various figures unless otherwise specified. For a better understanding, reference will be made to the following Detailed Description, which is to be read in association with the accompanying drawings, wherein:
FIGURE 1 illustrates a system environment in which various embodiments may be implemented;
FIGURE 2A shows a rack of blade servers that may be included in various embodiments; FIGURE 2B illustrates an embodiment of a blade server that may be included in a rack of blade servers such as that shown in FIGURE 2 A;
FIGURE 3 shows a mobile device that may be included in various embodiments;
FIGURE 4 illustrates a network device that may be included in various embodiments;
FIGURE 5 illustrates a system diagram showing an embodiment of a system that may be utilized to implement embodiments of processing, searching and managing data volumes;
FIGURE 6 illustrates a system diagram showing an alternative embodiment of a system that may be utilized to implement embodiments of processing, searching and managing data volumes;
FIGURE 7 illustrates a logical flow diagram generally showing one embodiment of an overview process for initiating a project that employs embodiments of the system diagrams shown in FIGURES 5 and/or 6;
FIGURE 8 illustrates a logical flow diagram generally showing one embodiment of a process for enabling access to a determined volume of data and restricting access to incoming data that exceeds the determined volume;
FIGURE 9 illustrates a logical flow diagram generally showing one embodiment of a process for managing an overflow index store;
FIGURE 10 illustrates a logical flow diagram generally showing one embodiment of a process for enabling a project owner to invite another user into the project;
FIGURE 11 illustrates a logical flow diagram generally showing one embodiment of an overview process for elastically scaling a project; FIGURE 12 illustrates a logical flow diagram generally showing one embodiment of a process for receiving resource allocation direction from an operator manipulating a user interface control;
FIGURE 13 illustrates a logical flow diagram generally showing one embodiment of a process for increasing a project's resources as part of elastic scaling;
FIGURE 14 depicts a logical flow diagram generally showing one embodiment of a process for responding to a requested decrease in project resources;
FIGURE 15 illustrates a logical flow diagram generally showing one embodiment of process for queuing resource requests;
FIGURE 16 depicts a logical flow diagram generally showing one embodiment of process for elastically scaling based on index storage size;
FIGURE 17 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on indexing performance;
FIGURE 18 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on an event time window;
FIGURE 19 illustrates a logical flow diagram generally showing one embodiment of a process for changing a number of indexers associated with an index store
FIGURE 20 shows a logical flow diagram generally showing one embodiment of a process to recall archived data for re -indexing;
FIGURE 21 illustrates a logical flow diagram generally showing one embodiment of a process for transitioning to scaled index stores;
FIGURE 22 illustrates a logical flow diagram generally showing one embodiment of an input proxy operation; and
FIGURE 23 illustrates a non-exhaustive example of a use case of an embodiment of user-interface controls for managing and processing data volumes.
DESCRIPTION OF VARIOUS EMBODIMENTS
Throughout the specification and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrase "in one embodiment" as used herein does not necessarily refer to the same embodiment, though it may. Furthermore, the phrase "in another embodiment" as used herein does not necessarily refer to a different embodiment, although it may. Thus, as described below, various embodiments may be readily combined, without departing from the scope or spirit of the invention.
In addition, as used herein, the term "or" is an inclusive "or" operator, and is equivalent to the term "and/or," unless the context clearly dictates otherwise. The term "based on" is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of "a," "an," and "the" include plural references. The meaning of "in" includes "in" and "on."
Various embodiments now will be described more fully hereinafter with reference to the accompanying drawings, which form a part hereof, and which show, by way of illustration, specific exemplary embodiments by which the invention may be practiced. The embodiments may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the embodiments to those skilled in the art. Among other things, the various embodiments may be methods, systems, media or devices. Accordingly, the various embodiments may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. The following detailed description is, therefore, not to be taken in a limiting sense.
For example embodiments, the following terms are also used herein according to the corresponding meaning, unless the context clearly dictates otherwise.
"Frontend User" as used herein may represent users at the outer cloud-based edge of some of the described embodiments. This outer edge may often be described herein as the Web Application Layer, or Cloud layer.
"Core User" as used herein may represent abstract users that are within the middle of the layers of described embodiments. These middle layers may often be described herein as Search Head layers, or Core layers.
"Apps" as used herein may collectively represent software programs and processes that may implement various features and processing for the various embodiments, some of which are described herein. Apps may give users insights into their projects with dashboards, reports, data inputs, and saved searches that may work in the project environment from the moment they are installed. Apps may include new views and dashboards that may completely reconfigure the way a project looks. Or, they may be as complex as an entirely new program. In some embodiments, Apps may also represent index resources. "Index", "Indexer", "Indexing", and "Index Storage" as used herein may represent elements of described embodiments that may index and store data and events. Indexers may collect, parse, and store data to facilitate fast and accurate information retrieval. Index design may incorporate interdisciplinary concepts from linguistics, cognitive psychology, mathematics, informatics, physics, and computer science. Also, Indexes may reside in flat files in a data store on a file system. Index files may be managed to facilitate flexible searching and fast data retrieval, eventually archiving them according to a configurable schedule. During indexing, incoming raw data may be processed to enable fast search and analysis, the results of which may be stored in an index. As part of the indexing process, the indexer may add knowledge to the data in various ways, including by: separating a data stream into individual, searchable events; creating or identifying timestamps; extracting fields such as host, source, and source type;
performing user-defined actions on the incoming data, such as identifying custom fields, masking sensitive data, writing new or modified keys, applying breaking rules for multi-line events, filtering unwanted events, and routing events to specified indexes or servers, and the like. In some embodiments, index store may store data indexed by one or more indexers. In at least one embodiment, indexer may include indexer severs.
"Roles" as used herein may represent definitions of the access permissions and authorizations that enable a Core User to execute various features, Apps, or access data, in the various embodiments. Roles may define how a Core User may interact with Apps, Indexes, Indexers, and the like. "Dimension" as used herein when describing elastic scaling, may refer to the particular kind of project resources that may be scaled, e.g., index storage, processing performance, time- window, price, network bandwidth, and the like. Embodiments may be arranged and configured to define dimensions differently, e.g., one embodiment may be arranged to define an elastic scaling dimension that represents network latency, where another embodiment may be arranged and configured to combine CPU quota, network latency, network bandwidth into a single dimension representing processing performance. One of ordinary skill in the art wi11 : ~L~ that there may be many combinations of resources that may arranged into elastic scaling dimensions.
Multi-tenant cloud-based data storage volumes may be built to support manual and rule- based elastic scaling. Elastic scaling may occur in one or more dimensions, such as scaling of storage, scaling of processing performance, or time windows of received data, and the like.
Elastic scaling may occur in one or more dimensions singly or in combination. If elastic scaling occurs, the appropriate objects, processes, memory, and storage may be scaled to meet the request resource increase or decrease. There may be numerous dependent or supporting systems that may need to scale in response to an elastic scaling request. For example, a request to elastically scale processing performance may require one or more elements of the system to be scaled in addition to the CPU quota, such as, disk quota, high local cache quota, and the like.
The following briefly describes embodiments in order to provide a basic understanding of some aspects of the invention. This brief description is not intended as an extensive overview. It is not intended to identify key or critical elements, or to delineate or otherwise narrow the scope. Its purpose is merely to present some concepts in a simplified form as a prelude to the more detailed description that is presented later.
Briefly stated, various embodiments are directed towards a system and method for a cloud-based front end that may abstract and enable access to the underlying cloud-hosted elements and objects that may be part of a multi-tenant application, such as a search application. Search objects may be employed to access indexed objects. Also, one or more projects can be elastically scaled for a user to provide resources that may meet the specific needs of each project.
In some embodiments, at least one project for at least one frontend user may be initialized. In at least one embodiment, the frontend user may be enabled to interact with the project through at least one application. Each project may have a corresponding index store, which may include one or more indexer servers. In at least one embodiment, a data volume limit may be established for the index store. The data volume limit may be based on a selection by the frontend user. Data provided for the project may be indexed. In some embodiments, if an amount of indexed data in the index store is less than the data volume limit, then the indexed data may be stored in the index store. In other embodiments, if the amount of indexed data in the index store at least equals the data volume limit, then the indexed data may be stored in an overflow index store. In at least one embodiment, the indexed data in the index store may be accessible to the frontend user and the indexed data in the overflow index store may be inaccessible to the frontend user (e.g., the frontend user may be prevented from accessing the data unless the user increases the data volume limit, such as by purchasing additional data storage).
If the data volume limit is increased, then at least a portion of indexed data in the overflow index store may be provided to the index store, such that the provided indexed data is accessible in the index store to the frontend user. In at least one embodiment, the data volume limit may be increased based on another selection by the frontend user.
In other embodiments, the user may be enabled to elastically scale one or more computing resources associated with a project. A resource allocation may be determined for indexing data provided for a project for each of a plurality of resources. In at least one embodiment, the resources may include at least one indexer server instance that indexes the provided data, a data store (i.e., storage capacity) that stores the indexed data, an event time window, or the like.
In at least one embodiment, a request may be provided from a user to scale at least one of the plurality of resources. In some embodiments, the user may be enabled to utilize a slider control (or other user interface) to manually adjust one or more computing resources. In response to the request, the resource allocation for at least one of the plurality of resources may be adjusted based on the request. At least one of the plurality of resources may be changed based on the adjusted resource allocation, which may be employed to subsequently index data provided for the project. In some embodiments, at least one type of additional resource may be allocated if the request indicates an increase in the plurality of resources. In other embodiments, less of at least one type of resource may be allocated if the change indicates a decrease in the plurality of resources.
In some embodiments, an increase or decrease in one resource may correspond to an increase or decrease in another resource. In at least one embodiment, if a resource is elastically scaled, then a number of indexers may be increased or decreased to accommodate the change in resources. In some embodiments, decreasing a number of indexers may include marking an indexer as read-only and increasing a number of indexers may include marking a read-only indexer as read/write and/or adding a new indexer. In some embodiments, changing resources may include changing a number of indexer servers employed to index data for the project, changing an amount of storage for the data store to store data for the project, changing an amount of memory employed to process the indexing of data for the project, changing an amount of processors employed to process the indexing of data for the project, or the like. Additionally, it should be appreciated that though many embodiments described herein may be cloud-based, embodiments should not be considered so limited. One of ordinary skill in the art will recognize that enabling embodiments may be arranged to be partially deployed in cloud-based configurations with some elements in the cloud and some elements separate from cloud based resources. Likewise, enabling embodiments may be arranged to be deployed and operate in configurations entirely separate from cloud-based resources.
Illustrative Operating Environment
FIGURE 1 shows components of an environment in which various embodiments may be practiced. Not all of the components may be required to practice the various embodiments, and variations in the arrangement and type of the components may be made without departing from the spirit or scope of the various embodiments.
In at least one embodiment, cloud network 102 enables one or more network services for a user based on the operation of corresponding arrangements 104 and 106 of virtually any type of networked computing device. As shown, the networked computing devices may include server network device 112, host network device 114, enclosure of blade servers 110, enclosure of server computers 116, super computer network device 118, and the like. Although not shown, one or more mobile devices may be included in cloud network 102 in one or more arrangements to provide one or more network services to a user. Also, these arrangements of networked computing devices may or may not be mutually exclusive of each other. Additionally, the user may employ a plurality of virtually any type of wired or wireless networked computing devices to communicate with cloud network 102 and access at least one of the network services enabled by one or more of arrangements 104 and 106. These networked computing devices may include tablet mobile device 122, handheld mobile device 124, wearable mobile device 126, desktop network device 120, and the like. Although not shown, in various embodiments, the user may also employ notebook computers, desktop computers,
microprocessor-based or programmable consumer electronics, network appliances, mobile telephones, smart telephones, pagers, radio frequency (RF) devices, infrared (IR) devices, Personal Digital Assistants (PDAs), televisions, integrated devices combining at least one of the preceding devices, and the like.
One embodiment of a mobile device is described in more detail below in conjunction with FIGURE 3. Generally, mobile devices may include virtually any substantially portable networked computing device capable of communicating over a wired, wireless, or some combination of wired and wireless network. In various embodiments, network 102 may employ virtually any form of communication technology and topology. For example, network 102 can include local area networks Personal Area Networks (PANs), (LANs), Campus Area Networks (CANs), Metropolitan Area Networks (MANs) Wide Area Networks (WANs), direct communication connections, and the like, or any combination thereof. On an interconnected set of LANs, including those based on differing architectures and protocols, a router acts as a link between LANs, enabling messages to be sent from one to another. In addition, communication links within networks may include virtually any type of link, e.g., twisted wire pair lines, optical fibers, open air lasers or coaxial cable, plain old telephone service (POTS), wave guides, acoustic, full or fractional dedicated digital communication lines including Tl , T2, T3, and T4, and/or other carrier and other wired media and wireless media. These carrier mechanisms may include E-carriers, Integrated Services
Digital Networks (ISDNs), universal serial bus (USB) ports, Firewire ports, Thunderbolt ports, Digital Subscriber Lines (DSLs), wireless links including satellite links, or other
communications links known to those skilled in the art. Moreover, these communication links may further employ any of a variety of digital signaling technologies, including without limit, for example, DS-0, DS-1, DS-2, DS-3, DS-4, OC-3, OC-12, OC-48, or the like. Furthermore, remotely located computing devices could be remotely connected to networks via a modem and a temporary communication link. In essence, network 102 may include virtually any communication technology by which information may travel between computing devices.
Additionally, in the various embodiments, the communicated information may include virtually any kind of information including, but not limited to processor-readable instructions, data structures, program modules, applications, raw data, control data, archived data, video data, voice data, image data, text data, and the like.
Network 102 may be partially or entirely embodied by one or more wireless networks. A wireless network may include any of a variety of wireless sub-networks that may further overlay stand-alone ad-hoc networks, and the like. Such sub-networks may include mesh networks, Wireless LAN (WLAN) networks, Wireless Router (WR) mesh, cellular networks, pico networks, PANs, Open Air Laser networks, Microwave networks, and the like. Network 102 may further include an autonomous system of intermediate network devices such as terminals, gateways, routers, switches, firewalls, load balancers, and the like, which are coupled to wired and/or wireless communication links. These autonomous devices may be operable to move freely and randomly and organize themselves arbitrarily, such that the topology of network 102 may change rapidly.
Network 102 may further employ a plurality of wired and wireless access technologies, e.g., 2nd (2G), 3rd (3G), 4th (4G), 5th (5G) generation wireless access technologies, and the like, for mobile devices. These wired and wireless access technologies may also include Global System for Mobile communication (GSM), General Packet Radio Services (GPRS), Enhanced Data GSM Environment (EDGE), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution Advanced (LTE), Universal Mobile Telecommunications System (UMTS), Orthogonal frequency-division multiplexing (OFDM), Wideband Code Division Multiple Access (W-CDMA), Code Division Multiple Access 2000 (CDMA2000), Evolution-Data Optimized (EV-DO), High-Speed Downlink Packet Access (HSDPA), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), ultra wide band (UWB), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), any portion of the Open Systems Interconnection (OSI) model protocols, Short Message Service (SMS), Multimedia Messaging Service (MMS), Web Access Protocol (WAP), Session Initiation Protocol/Real-time Transport Protocol (SIP/RTP), or any of a variety of other wireless or wired communication protocols. In one non- limiting example, network 102 may enable a mobile device to wirelessly access a network service through a combination of several radio network access technologies such as GSM, EDGE, SMS, HSDPA, and the like.
Enclosure of Blade Servers
FIGURE 2A shows one embodiment of an enclosure of blade servers 200, which are also illustrated in FIGURE 1. Enclosure of blade servers 200 may include many more or fewer components than those shown in FIGURE 2A. However, the components shown are sufficient to disclose an illustrative embodiment. Generally, a blade server is a stripped down server computing device with a modular design optimized to minimize the use of physical space and energy. A blade enclosure can include several blade servers and provide each with power, cooling, network interfaces, input/output interfaces, and resource management. Although not shown, an enclosure of server computers typically includes several computers that merely require a network connection and a power cord connection to operate. Each server computer often includes redundant components for power and interfaces.
As shown in the figure, enclosure 200 contains power supply 204, and input/output interface 206, rack logic 208, several blade servers 210, 212, 214, and 216, and backplane 202. Power supply 204 provides power to each component and blade server within the enclosure. The input/output interface 206 provides internal and external communication for components and blade servers within the enclosure. Backplane 208 can enable passive and active communication of power, logic, input signals, and output signals for each blade server. Illustrative Blade Server
FIGURE 2B illustrates an illustrative embodiment of blade server 250, which may include many more or fewer components than those shown. As shown in FIGURE 2A, a plurality of blade servers may be included in one enclosure that shares resources provided by the enclosure to reduce size, power, and cost. Blade server 250 includes processor 252 which communicates with memory 256 via bus
254. Blade server 250 also includes input/output interface 290, processor-readable stationary storage device 292, and processor-readable removable storage device 294. Input/output interface 290 can enable blade server 250 to communicate with other blade servers, mobile devices, network devices, and the like. Interface 190 may provide wireless and/or wired communication links for blade server. Processor-readable stationary storage device 292 may include devices such as an electromagnetic storage device (hard disk), solid state hard disk (SSD), hybrid of both an SSD and a hard disk, and the like. Also, processor-readable removable storage device 294 enables processor 252 to read non-transitive storage media for storing and accessing processor- readable instructions, modules, data structures, and other forms of data. The non-transitive storage media may include Flash drives, tape media, floppy media, and the like.
Memory 256 may include Random Access Memory (RAM), Read-Only Memory (ROM), hybrid of RAM and ROM, and the like. As shown, memory 256 includes operating system 258 and basic input/output system (BIOS) 260 for enabling the operation of blade server 250. In various embodiments, a general-purpose operating system may be employed such as a version of UNIX, or LINUX™, or a specialized server operating system such as Microsoft's Windows Server™ and Apple Computer's IoS Server™.
Memory 256 further includes one or more data storage 270, which can be utilized by blade server 250 to store, among other things, applications 280 and/or other data. Data stores 270 may include program code, data, algorithms, and the like, for use by processor 252 to execute and perform actions. In one embodiment, at least some of data store 270 might also be stored on another component of blade server 250, including, but not limited to, processor- readable removable storage device 294, processor-readable stationary storage device 292, or any other processor-readable storage device (not shown). Data storage 270 may include, for example, raw data 272, and indexed data 274.
Applications 280 may include processor executable instructions which, when executed by blade server 250, transmit, receive, and/or otherwise process messages, audio, video, and enable communication with other networked computing devices. Examples of application programs include database servers, file servers, calendars, transcoders, and so forth.
Applications 280 may include, for example, cloud search application 282, and indexing application 284.
Human interface components (not pictured), may be remotely associated with blade server 250, which can enable remote input to and/or output from blade server 250. For example, information to a display or from a keyboard can be routed through the input/output interface 290 to appropriate peripheral human interface components that are remotely located. Examples of peripheral human interface components include, but are not limited to, an audio interface, a display, keypad, pointing device, touch interface, and the like.
Illustrative Mobile Device
FIGURE 3 shows one embodiment of mobile device 300 that may include many more or less components than those shown. Mobile device 300 may represent, for example, at least one embodiment of mobile devices shown in FIGURE 1.
Mobile device 300 includes processor 302 in communication with memory 304 via bus 328. Mobile device 300 also includes power supply 330, network interface 332, audio interface 356, display 350, keypad 352, illuminator 354, video interface 342, input/output interface 338, haptic interface 364, global positioning systems (GPS) receiver 358, Open air gesture interface 360, temperature interface 362, camera(s) 340, projector 346, pointing device interface 366, processor-readable stationary storage device 334, and processor-readable removable storage device 336. Power supply 330 provides power to mobile device 300. A rechargeable or non- rechargeable battery may be used to provide power. The power may also be provided by an external power source, such as an AC adapter or a powered docking cradle that supplements and/or recharges the battery. And in one embodiment, although not shown, a gyroscope may be employed within mobile device 300 to measuring and/or maintaining an orientation of mobile device 300.
Mobile device 300 may optionally communicate with a base station (not shown), or directly with another computing device. Network interface 332 includes circuitry for coupling mobile device 300 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the Open Systems Interconnection (OSI) model for mobile communication (GSM), code division multiple access (CDMA), time division multiple access (TDMA), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), Short Message Service (SMS), Multimedia Messaging Service (MMS), general packet radio service (GPRS), WAP, ultra wide band (UWB), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), Session Initiation Protocol/Real-time Transport Protocol (SIP/RTP), or any of a variety of other wireless communication protocols. Network interface 332 is sometimes known as a transceiver, transceiving device, or network interface card (NIC). Audio interface 356 is arranged to produce and receive audio signals such as the sound of a human voice. For example, audio interface 356 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others and/or generate an audio
acknowledgement for some action. A microphone in audio interface 356 can also be used for input to or control of mobile device 300, e.g., using voice recognition, detecting touch based on sound, and the like.
Display 350 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computing device. Display 350 may also include a touch interface 344 arranged to receive input from an object such as a stylus or a digit from a human hand, and may use resistive, capacitive, surface acoustic wave (SAW), infrared, radar, or other technologies to sense touch and/or gestures. Projector 346 may be a remote handheld projector or an integrated projector that is capable of projecting an image on a remote wall or any other reflective object such as a remote screen. Video interface 342 may be arranged to capture video images, such as a still photo, a video segment, an infrared video, or the like. For example, video interface 342 may be coupled to a digital video camera, a web-camera, or the like. Video interface 342 may comprise a lens, an image sensor, and other electronics. Image sensors may include a complementary metal-oxide- semiconductor (CMOS) integrated circuit, charge-coupled device (CCD), or any other integrated circuit for sensing light.
Keypad 352 may comprise any input device arranged to receive input from a user. For example, keypad 352 may include a push button numeric dial, or a keyboard. Keypad 352 may also include command buttons that are associated with selecting and sending images. Illuminator 354 may provide a status indication and/or provide light. Illuminator 354 may remain active for specific periods of time or in response to events. For example, when illuminator 354 is active, it may backlight the buttons on keypad 352 and stay on while the mobile device is powered. Also, illuminator 354 may backlight these buttons in various patterns when particular actions are performed, such as dialing another mobile device. Illuminator 354 may also cause light sources positioned within a transparent or translucent case of the mobile device to illuminate in response to actions. Mobile device 300 also comprises input/output interface 338 for communicating with external peripheral devices or other computing devices such as other mobile devices and network devices. The peripheral devices may include an audio headset, display screen glasses, remote speaker system, remote speaker and microphone system, and the like. Input/output interface 338 can utilize one or more technologies, such as Universal Serial Bus (USB), Infrared, WiFi, WiMax, Bluetooth™, and the like. Haptic interface 364 is arranged to provide tactile feedback to a user of the mobile device. For example, the haptic interface 364 may be employed to vibrate mobile device 300 in a particular way when another user of a computing device is calling. Temperature interface 362 may be used to provide a temperature measurement input and/or a temperature changing output to a user of mobile device 300. Open air gesture interface 360 may sense physical gestures of a user of mobile device 300, for example, by using single or stereo video cameras, radar, a gyroscopic sensor inside a device held or worn by the user, or the like. Camera 340 may be used to track physical eye movements of a user of mobile device 300.
GPS transceiver 358 can determine the physical coordinates of mobile device 300 on the surface of the Earth, which typically outputs a location as latitude and longitude values. GPS transceiver 358 can also employ other geo-positioning mechanisms, including, but not limited to, triangulation, assisted GPS (AGPS), Enhanced Observed Time Difference (E-OTD), Cell Identifier (CI), Service Area Identifier (SAI), Enhanced Timing Advance (ETA), Base Station Subsystem (BSS), or the like, to further determine the physical location of mobile device 300 on the surface of the Earth. It is understood that under different conditions, GPS transceiver 358 can determine a physical location for mobile device 300. In at least one embodiment, however, mobile device 300 may, through other components, provide other information that may be employed to determine a physical location of the device, including for example, a Media Access Control (MAC) address, IP address, and the like. Human interface components can be peripheral devices that are physically separate from mobile device 300, allowing for remote input and/or output to mobile device 300. For example, information routed as described here through human interface components such as display 350 or keyboard 352 can instead be routed through network interface 332 to appropriate human interface components located remotely. Examples of human interface peripheral components that may be remote include, but are not limited to, audio devices, pointing devices, keypads, displays, cameras, projectors, and the like. These peripheral components may communicate over a Pico Network such as Bluetooth™, Zigbee™ and the like. One non-limiting example of a mobile device with such peripheral human interface components is a wearable computing device, which might include a remote pico projector along with one or more cameras that remotely communicate with a separately located mobile device to sense a user's gestures toward portions of an image projected by the pico projector onto a reflected surface such as a wall or the user's hand.
A mobile device may include a browser application that is configured to receive and to send web pages, web-based messages, graphics, text, multimedia, and the like. The mobile device's browser application may employ virtually any programming language, including a wireless application protocol messages (WAP), and the like. In at least one embodiment, the browser application is enabled to employ Handheld Device Markup Language (HDML), Wireless Markup Language (WML), WMLScript, JavaScript, Standard Generalized Markup Language (SGML), HyperText Markup Language (HTML), extensible Markup Language (XML), HTML5, and the like.
Memory 304 may include Random Access Memory (RAM), Read-Only Memory (ROM), and/or other types of memory. Memory 304 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data. Memory 304 stores a basic input/output system (BIOS) 308 for controlling low-level operation of mobile device 300. The memory also stores an operating system 306 for controlling the operation of mobile device 300. It will be appreciated that this component may include a general-purpose operating system such as a version of UNIX, or LINUX™, or a specialized mobile computer communication operating system such as Windows Mobile™, or the Symbian® operating system. The operating system may include, or interface with a Java virtual machine module that enables control of hardware components and/or operating system operations via Java application programs.
Memory 304 further includes one or more data storage 310, which can be utilized by mobile device 300 to store, among other things, applications 320 and/or other data. For example, data storage 310 may also be employed to store information that describes various capabilities of mobile device 300. The information may then be provided to another device based on any of a variety of events, including being sent as part of a header during a communication, sent upon request, or the like. Data storage 310 may also be employed to store social network"'" ~ information including address books, buddy lists, aliases, user profile information, or the like. Data storage 310 may further include program code, data, algorithms, and the like, for use by a processor, such as processor 302 to execute and perform actions. In one embodiment, at least some of data storage 310 might also be stored on another component of mobile device 300, including, but not limited to, non-transitory processor-readable removable storage device 336, processor-readable stationary storage device 334, or even external to the mobile device. Data storage 310 may include, for example, raw data 312, and index data 314.
Applications 320 may include computer executable instructions which, when executed by mobile device 300, transmit, receive, and/or otherwise process instructions and data.
Applications 320 may include, for example, Cloud Search application 322, and Indexing
Application 324. Other examples of application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth.
Illustrative Network Device
FIGURE 4 shows one embodiment of network device 400 that may be included in a system implementing the invention. Network device 400 may include many more or less components than those shown in FIGURE 4. However, the components shown are sufficient to disclose an illustrative embodiment for practicing the present invention. Network device 400 may represent, for example, one embodiment of at least one of network device 112, 114, or 120 of FIGURE 1.
As shown in the figure, network device 400 includes a processor 402 in communication with a memory 404 via a bus 428. Network device 400 also includes a power supply 430, network interface 432, audio interface 456, display 450, keyboard 452, input/output interface
438, processor-readable stationary storage device 434, and processor-readable removable storage device 436. Power supply 430 provides power to network device 400.
Network interface 432 includes circuitry for coupling network device 400 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the Open Systems Interconnection model (OSI model), global system for mobile
communication (GSM), code division multiple access (CDMA), time division multiple access (TDMA), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), Short Message Service (SMS), Multimedia Messaging Service (MMS), general packet radio service (GPRS), WAP, ultra wide band (UWB), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), Session Initiation Protocol/Real-time Transport Protocol (SIP/RTP), or any of a variety of other wired and wireless communication protocols. Network interface 432 is sometimes known as a transceiver, transceiving device, or network interface card (NIC). Network device 400 may optionally communicate with a base station (not shown), or directly with another computing device.
Audio interface 456 is arranged to produce and receive audio signals such as the sound of a human voice. For example, audio interface 456 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others and/or generate an audio
acknowledgement for some action. A microphone in audio interface 456 can also be used for input to or control of network device 400, for example, using voice recognition.
Display 450 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computing device. Display 450 may be a handheld projector or pico projector capable of projecting an image on a wall or other object.
Network device 400 also may also comprise input/output interface 438 for
communicating with external devices not shown in FIGURE 4. Input/output interface 438 can utilize one or more wired or wireless communication technologies, such as USB™, Firewire™, WiFi, WiMax, Thunderbolt™, Infrared, Bluetooth™, Zigbee™, serial port, parallel port, and the like.
Human interface components can be physically separate from network device 400, allowing for remote input and/or output to network device 400. For example, information routed as described here through human interface components such as display 450 or keyboard 452 can instead be routed through the network interface 432 to appropriate human interface components located elsewhere on the network. Human interface components include any component that allows the computer to take input from, or send output to, a human user of a computer.
Memory 404 may include Random Access Memory (RAM), Read-Only Memory (ROM), and/or other types of memory. Memory 404 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data. Memory 404 stores a basic input/output system (BIOS) 408 for controlling low-level operation of network device 400. The memory also stores an operating system 406 for controlling the operation of network device 400. It will be appreciated that this component may include a general-purpose operating system such as a version of UNIX, or LINUX™, or a specialized operating system such as Microsoft
Corporation's Windows ® operating system, or the Apple Corporation's iOS® operating system. The operating system may include, or interface with a Java virtual machine module that enables control of hardware components and/or operating system operations via Java application programs. Memory 404 further includes one or more data storage 410, which can be utilized by network device 400 to store, among other things, applications 420 and/or other data. For example, data storage 410 may also be employed to store information that describes various capabilities of network device 400. The information may then be provided to another device based on any of a variety of events, including being sent as part of a header during a
communication, sent upon request, or the like. Data storage 410 may also be employed to store social networking information including address books, buddy lists, aliases, user profile information, or the like. Data stores 410 may further include program code, data, algorithms, and the like, for use by a processor, such as processor 402 to execute and perform actions. In one embodiment, at least some of data store 410 might also be stored on another component of network device 400, including, but not limited to, non-transitory media inside processor- readable removable storage device 436, processor-readable stationary storage device 434, or any other computer-readable storage device within network device 400, or even external to network device 400. Data storage 410 may include, for example, raw data 412, and indexed data 414.
Applications 420 may include computer executable instructions which, when executed by network device 400, transmit, receive, and/or otherwise process messages (e.g., SMS,
Multimedia Messaging Service (MMS), Instant Message (IM), email, and/or other— ~~ audio, video, and enable telecommunication with another user of another mobile device. Other examples of application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth. Applications 420 may include, for example, cloud search application 422, and indexing application 424.
Illustrative System Diagram
FIGURE 5 depicts an embodiment where the top layer of the figure, indicated as Web Application Layer 502, may represent a cloud-based front end that may abstract and enable access to the underlying cloud-hosted elements and objects that may be part of a multi-tenant application that may comprise Core User (e.g., Core User 518 and Core User 520), Roles (e.g., Role 522 and Role 524), and Apps (e.g., App 526 and App 528), indicated as the Search Head 504. The objects in the Search Head 504 layer may be employed to access the Index objects located in the Indexers 506 layer of FIGURE 5. The two projects, Project 1 (i.e., project 508) and Project 2 (i.e., project 510) in the Web Application Layer 502 may be examples of projects that have been elastically scaled to provide resources that may meet the specific needs of each project. As depicted, Project 1 may require twice as much performance and Project 2, the figure represents this with the additional Apps (i.e., App 526 and App 528) in the Search Head 504 layer, and the two index stores (i.e., index 530 and index 532) in the Indexer 506 layer.
Also, the embodiment in FIGURE 5, demonstrates that a Frontend User may have access to one or more Projects. For example, frontend user 512 may have access to project 508, Frontend User 516 may have access to project 510, and Frontend User 514 may have access to both project 508 and project 510. An embodiment may enable the Search Head 504 layer of a Project (e.g., project 508) to create unique Core Users in the Search Head 504 layer that may be linked to Frontend Users from the Web Application Layer 502. For example, Core User 518 may be linked to Frontend User 512 and Core User 520 may be linked to Frontend User 514.
A Role of the Core User that may be linked to the Frontend User may enable access control for the Apps and other objects within the Search Head layer. For example, Role 522 of Core User 518 may enable Frontend User 512 to access control for App 526 and App 528. Similarly, Role 524 of Core User 520 may enable Frontend User 514 to access control for App 526 and App 528. The Role may determine the specific features that are made available to a Core User, and a Frontend User may have its access within a Project controlled by the Role of the Core User it is linked with. In other words, as depicted by FIGURE 5, the Roles may be thought of as providing pathways for the Frontend User to reach the Apps. In some
embodiments, the Roles may also control access to indexers. For example, Role 522 may control access of Core User 518 to Apps 526 and/or 528 as well as index 530 and/or 532.
FIGURE 6, depicts an embodiment where Frontend User 1 (i.e., Frontend User 612) may be the owner of two projects, Project 1 (i.e., Project 608) and Project 2 (i.e., Project 610). As shown in the figure, Frontend User 2 (i.e., Frontend User 614) may not be an owner of a project. But, Frontend User 2 illustrates how a Frontend User may be invited by a project owner, represented in this embodiment by the depicted Frontend User 1. In at least one embodiment, Frontend User 612 and/or Frontend User 614 may be an embodiment of a Frontend User from FIGURE 5. In some embodiments, Project 608 and Project 610 may be embodiments of Project 508 and Project 510 of FIGURE 5, respectively. Web Application Layer 602, Search Head 604, and Indexers 606 may be embodiments of Web Application Layer 502, Search Head 504, and Indexers 506 of FIGURE 5, respectively.
Similar to the embodiment depicted in FIGURE 5, this embodiment may enable unique Core Users to be created and linked to Frontend Users. For example, Core User 618 and Core User 620 may be linked to Frontend User 612 and Core User 622 may be linked to Frontend User 614. For each unique Core User, a Role may be created to define the features that may be accessed by the Frontend User. For example, Role 624 of Core User 618 may define the features that may be accessed by Frontend User 612 (e.g., App 626).
The embodiment depicted in FIGURE 6, may also enable Indexes (e.g., Index 632), Ports (e.g., Port 634), and associated storage to be located separately from the Search Head/Core layer objects. Also, embodiments may locate Indexes, and Ports across more than one computing device e.g., computing device 628 and computing device 630), including, network computing devices, blade servers, individual blade computers, cloud-based compute instances, virtual machines, and the like. Further, Indexes that may be associated with different Projects may be co-located on the same computing device, such as, network computing devices, blade servers, individual blade computers, cloud-based compute instances, virtual machines, and the like. General Operation of Data Volume Management
The operation of certain aspects of the invention will now be described with respect to FIGURES 7-10. FIGURE 7 illustrates a logical flow diagram generally showing one
embodiment of an overview process for initiating a project that employs embodiments of the system diagrams shown in FIGURES 5 and/or 6. In some embodiments, process 700 of FIGURE 7 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 700 or portions of process 700 of FIGURE 7 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 700 or portions of process 700 of FIGURE 7 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Process 700 begins, after a start block, at block 702 where at least one project may be initialized for at least one Frontend User. In some embodiments, Roles, Apps, Indexes, or the like may be created and/or configured for each unique project. In some embodiments, blocks 704, 706, 708, 710, and 712 may be performed in an order other than illustrated in FIGURE 7.
Processing continues at block 704, where one or more Apps may be created and/or associated with the project. In some embodiments, the associated Apps may be identified by the Frontend User.
Process 700 continues at block 706, where at least one Role may be determined for the project. In some embodiments, an operator and/or the Frontend User may generate and/or modify the access permissions associated with the Role. In at least one embodiment, at least one Role for each frontend user may be determined that may establish permissions for interacting with the at least one project and the at least one application for each Frontend User. In at least one embodiment, the Roles may configure access control for a Frontend User to the Apps and/or the indexers. For example, in some embodiments, Roles may enable one Frontend User to access one App (and/or indexer ), but not another App (and/or indexer). In some embodiments, a plurality of Roles may be determined for each Core User.
Process 700 proceeds to block 708, where a Core User on the project may be created for the Frontend User. In at least one embodiment, creating a Core User may include determining a Core User that corresponds to a frontend user and a project, wherein the Core User enables the Frontend User to interact with the project. In some embodiments, the Core User may inherit one or more corresponding Roles determined at block 706.
Process 700 continues next at block 710, where the Frontend User may be linked with the Core User and the Role.
Process 700 proceeds to block 712, where an available index store may be established for the project. In at least one embodiment, an available index store may refer to an index store that is accessible by the Frontend User. In some embodiments, the Frontend User may be enabled to search the available index store. In at least one embodiment, the available index store may include one or more indexer instances. In some embodiments, the indexers may be configured based on the Roles. As described above, access to the index store (and/or the Apps) may be controlled by the Roles. In some other embodiments, a port may be assigned to each individual index.
Continuing to block 714, the index store may be dynamically scaled, which is described in more detail below in conjunction with FIGURE 8. Briefly, however, the index store may be dynamically scaled based on the Frontend User account. For example, if the index store exceeds a pre -purchased data volume, then incoming data may continue to be index, but the Frontend User may not be able to access the incoming data until the Frontend User pays for an increase in the data volume.
In some embodiments, the Frontend User may be enabled to select when and/or how dynamic scaling may be employed. For example, in some embodiments, the Frontend User may determine that if an index size limit is reached, at least one of the following may occur, including: automatically deleting additional data provided for the project beyond the size limit, the indexers may automatically stop indexing data, automatically increasing the storage size of data supported for a project, or the like.
After block 714, process 700 may return control to a calling process.
FIGURE 8 illustrates a logical flow diagram generally showing one embodiment of a process for enabling access to a determined volume of data and restricting access to incoming data that exceeds the determined volume. In some embodiments, process 800 of FIGURE 8 mav be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 800 or portions of process 800 of FIGURE 8 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 800 or portions of process 800 of FIGURE 8 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Process 800 begins, after a start block, at block 802, data may be provided. In some embodiments, the data may be incoming data for a given project. In other embodiments, the data may be archived data for a given project that may be recovered.
Process 800 continues at block 804, where the data may be indexed.
Process 800 proceeds to decision block 806, where a determination may be made whether a data volume limit is exceeded by the indexed data. In one embodiment, each project may include a data volume limit. In another embodiment, each frontend user (or group of frontend users) may be associated with a data volume limit. In at least one embodiment, the total stored data plus the indexed data may be compared to a maximum allowable data volume limit. In some embodiment, the data volume limit may be based on an amount of storage space purchased by a user (e.g., a Frontend User). If the data volume limit is exceeded, then process 800 may flow to block 810; otherwise, process 800 may flow to block 808.
At block 808, the indexed data may be stored in an available index store. In some embodiments, an available index store may refer to an index store that is accessible to the user. Process 800 may proceed to block 818.
If, at decision block 806, the data volume limit is exceeded, then process 800 may flow from decision block 806 to block 810. At block 810, the indexed data may be stored in an overflow index store. In some embodiments, the overflow index store may be unavailable to the user. The overflow index store allows continued indexing of a user's data even if the amount of data has exceeded a purchased maximum data volume. However, the user may not have access to the data in the overflow index store, unless the user purchases additional data volume.
In at least one embodiment, the overflow index store may be a FIFO buffer. The data in the overflow index store may remain in the overflow index store until a user purchases additional data volume and/or the data is removed (e.g., deleted/archived) after a predetermined amount of time. Some embodiments for removing data from the overflow index store are described in more detail below in conjunction with FIGURE 9.
In any event, process 800 next flows to decision block 812, where a determination may be made whether access to data in the overflow index store is enabled. In some embodiments, access may be enabled if the user increases the data storage volume, such as by purchasing additional data storage. In at least one embodiment, the data storage volume may be elastically scaled as described in more detail below in conjunction with FIGURES 11-23. If access to the overflow index store is enabled, process 800 may flow to block 814; otherwise, process 800 may flow to block 818.
At block 814, an amount of data in the overflow index store to make available may be determined. In some embodiments, the amount of data may be equal to or less than an increase in data storage volume. For example, if the user purchases an extra 2 GB of data storage, then it may be determined that up to 2 GB of data from the overflow index store may be made available. In some embodiments, the determined amount of data may be pulled from the overflow index store based on a FIFO (first in first out) buffer architecture of the overflow index store. In some embodiment, the indexed data that exceeds the data volume limit of the available index store is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit. Process 800 proceeds to block 816, where the determined amount of data may be provided to the available index store. In at least one embodiment, determining an amount of indexed data to copy from the overflow index store to the index store based on at least an increase to the data volume limit. In another embodiment, providing the data from the overflow index store to the available index store may include modifying an index store flag; copying the data from the overflow index store to the available index store, or the like. In some
embodiments, at least a portion of the indexed data in the overflow index store may be copied to archival storage. For example, the indexed data provided from the overflow index store to the available index store may also be provided to and/or copied to archival storage. Process 800 continues next at block 818, where the user may be enabled to access the available index store. In at least one embodiment, the user may be enabled to perform searches for data in the available index store.
After block 800, process 800 may return control to a calling process.
FIGURE 9 illustrates a logical flow diagram generally showing one embodiment of a process for managing an overflow index store. In some embodiments, process 900 of FIGURE 9 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 900 or portions of process 900 of FIGURE 9 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 900 or portions of process 900 of FIGURE 9 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Process 900 begins, after a start block, at decision block 902, where a determination may be made whether to remove data from the overflow index store. In some embodiments, data may be removed from the overflow index store when it is moved to the available index store. In other embodiments, data may be removed from the overflow index store after a predetermined amount of time. In yet other embodiments, data may be removed from the overflow index store if the amount of data in overflow index store exceeds a maximum overflow index store volume. If data is removed from the overflow index store, then process 900 may flow to decision block 904; otherwise, process 900 may return control to a calling process.
At decision block 904, a determination may be made whether the removed data is moved to archival storage. In some embodiments, this determination may be based on user preferences, settings, rules (e.g., Roles), or the like. For example, an account for the user may indicate that the data removed from the overflow index store may be deleted without archiving the data. In other embodiments, the determination may be based on whether the data was added to the available index store at block 816 of FIGURE 8. In one such embodiment, if the data was added to the available index store, then the data may be moved to archival storage. If the data is moved to archival storage, then process 900 may flow to block 906; otherwise, process 900 may flow to block 908. At block 906, data may be moved from the overflow index store to archival storage. In one embodiment, data may be moved to the archival storage based on a FIFO buffer architecture of the overflow index store. In some embodiments, an amount of data to be moved to the archival storage may be based on the FIFO buffer. In other embodiments, the amount of data to be moved may be based on an amount of data added to the available index store at block 816 of FIGURE 8.
Process 900 proceeds to block 908, where data may be deleted from the overflow index store. In one embodiment, data may be deleted based on a FIFO buffer architecture of the overflow index store. In some embodiments, an amount of data to be deleted may be based on the FIFO buffer. In other embodiments, the amount of data to be deleted may be based on an amount of data added to the available index store at block 816 of FIGURE 8.
After block 908, process 900 may return control to a calling process.
FIGURE 10, illustrates a logical flow diagram generally showing one embodiment of a process for enabling a project owner to invite another user into the project. In some
embodiments, process 1000 of FIGURE 10 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1000 or portions of process 1000 of FIGURE 10 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
embodiments, process 1000 or portions of process 1000 of FIGURE 10 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Process 1000 begins, after a start block, at block 1002, where a project owner may be enabled to send an invitation to an invited user. Also, any project user, or administrative user having the authority to invite users to a particular project may send invitations. Further, though not depicted in FIGURE 10, administrators, or other authorized users may define negative or positive restrictions on the users that may be invited to project. Restrictions may be defined for many purposes including, project security, privacy, licensing, and the like. Restrictions may be based on individual users as well as groups, lists, patterns, and the like.
Process 1000 proceeds to decision block 1004, where a determination may be made whether a user accepted the invitation. In at least one embodiment, an invited user may have the choice to accept the invitation. However, some embodiments may automatically join invited users to a project without giving them the option to decline the invitation. If an invited user declines or ignores an invitation, then process 1000 may return to the calling process; otherwise, process 1000 may flow to decision block 1006. In some embodiments the sender of the invitation may receive notice that the invitation has been declined or ignored.
At decision block 1006, a determination may be made whether the invited user has a Frontend User account. If the invited user has an active Frontend User account, then process 1000 may flow to block 1010; otherwise, process 1000 may flow to block 1008.
At block 1008, the invited user may create a Frontend User account.
Processing continues at block 1010, where a new Core User on a project may be creased. Proceeding to block 1012, a corresponding new Role for the new Core User may also be determined. In some embodiments, one or more new Roles may be created for the new Core User. In other embodiments, the new Core User may inherit one or more Roles from the user who initiated in invitation.
Process 1000 proceeds to block 1014, where the invited Frontend User may be linked with the new Core User and the new Role. After block 1014, process 1000 may return control to a calling process.
General Operation of Elastic Scaling of Data Volumes
The operation of certain aspects of the invention will now be described with respect to FIGURES 11-22. FIGURE 11 illustrates a logical flow diagram generally showing one embodiment of an overview process for elastically scaling a project. In some embodiments, process 1100 of FIGURE 11 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1100 or portions of process 1100 of FIGURE 11 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1100 or portions of process 1100 of FIGURE 11 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. Process 1100 begins, after a start block, at block 1102, where resources allocated for use by the Project may be determined to be sub-optimal. Such a determination may be made by many methods, including, rule -based systems, or manual systems. Embodiments may apply various rule-based systems that may enable determining that a project is operating with suboptimal resources. Such rule-based systems may include, detecting if a performance metric threshold has been exceeded, triggering resource allocations based on schedule, triggering resource allocations based on pattern prediction that may indicate a potential change in utilization requirements, and the like. In addition, an operator may determine by observation that resources are suboptimal.
Process 1100 next proceeds to block 1104, where, the optimal level of project resources may be determined. In some embodiments, optimal resources may be static or dynamic. In at least one embodiment, optimal resources may be automatically determined by rule-based systems, predictive computing techniques, or the like. In another embodiment, optimal resources may be manually determined by an operator. The specific allocation of resources may be determined in part by the characteristics of the project, or they may be determined by factors external to project. For example, a project that requires a very large dataset may have an optimal resource allocation that emphasizes increased data storage volume over processor performance.
Process 1100 continues at block 1106, where the project resources may be scaled to support the determined optimal level. Additional non-limiting embodiments of elastic scaling of project resources are described below. After block 1106, process 1100 may return control to a calling process.
FIGURE 12 illustrates a logical flow diagram generally showing one embodiment of a process for receiving resource allocation direction from an operator manipulating a user interface control. In some embodiments, process 1200 of FIGURE 12 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1200 or portions of process 1200 of FIGURE 12 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1200 or portions of process 1200 of FIGURE 12 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. The user interface control may have a user interface that represents a slider, dial, selection lists, radio buttons, check boxes, and the like. Embodiments may be arranged to map various elastic scaling dimensions to a control. For example, a control may represent processing performance, index storage, time-window of received data, and the like. An embodiment of a use case of a slider user interface is described in more detail below in conjunction with FIGURE 23.
Process 1200 begins, after a start block, at block 1202, where a control change (delta) may be received. In at least one embodiment, an operator may change, or manipulate the control to provide the delta. In some embodiments, a signal, message, or notification from the control may be received indicating that a change in the control may have occurred (i.e., the delta).
Process 1200 proceeds next to decision block 1204, where a determination may be made whether the delta is greater than zero. In at least one embodiment, a delta greater than zero may indicate a request to increase resources for the instant project. If the delta is greater than zero, then process 1200 may flow to block 1206; otherwise, process 1200 may flow to decision block 1208.
At block 1206, the resources for the project may be increased. One embodiment of a process for increasing resources for a project is described in more detail below in conjunction with FIGURE 13. After block 1206, process 1200 may return control to a calling process.
If, at decision block 1204, the delta is not greater than zero, then processing may flow from decision block 1204 to decision block 1208. At decision block 1208 a determination may be made whether the delta is less than zero. In at least one embodiment, a delta less than zero may indicate a request to decrease resources for the instant project. If the delta is less than zero, then process 1200 may flow to block 1210; otherwise, process 1200 may return control to a calling process.
At block 1210, the resources for the project may be decreased. One embodiment of a process for decreasing resources for a project is described in more detail below in conjunction with FIGURE 10.
Note, the concept of increasing or decreasing resources as used herein is generalized to refer to increasing or decreasing project resources in a defined and configured "dimension." Thus, it is important to recognize that increasing project resources in a particular defined dimension may actually require some computing resources to be increased while other may be decreased. For example, if for an embodiment, a single project resource elastic scaling dimension has been arranged such the increasing a slider (delta > zero) means to emphasize storage and de-emphasize processing speed, then the effect may be to increase storage and reduce the CPU quota. Whereas, staying with this example, decreasing the slider (delta < zero) may mean to de-emphasize storage and emphasize processing speed, then the effect may be to de-increase storage and increase the CPU quota. In some embodiments there may be a separate slider or control for many of the computing and performance attributes that make up project resources.
One of ordinary skill in the art will appreciate that sliders and controls may be arranged and configured in many ways to provide operators with a user-interface for controlling the elastic scaling of project resources. Embodiments may range from single controls that effect many elements of the computing process with a single action, to many fine grain controls that may be designed to elastically scale individual computing performance attributes. For example, an embodiment enabling fine grain control may have separate controls to adjust CPU quota, RAM use, disk access, network bandwidth, process priority, and the like. Similarly, on the other end of the spectrum, an embodiment may implement a single high level control in the aggregate that may be arranged to hide and abstract the separate adjustments to CPU quota, RAM use, disk access, network bandwidth, process priority, and the like. In the case of high-level controls the operator may not be required to have knowledge that a single change in the high level control may affect many of the project's resource components.
Next, process 1200 may return control to a calling process.
FIGURE 13 illustrates a logical flow diagram generally showing one embodiment of a process for increasing a project's resources as part of elastic scaling. In some embodiments, process 1300 of FIGURE 13 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1300 or portions of process 1300 of FIGURE 13 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1300 or portions of process 1300 of FIGURE 13 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. Process 1300 begins, after a start block, at block 1302, where particular computing resources that correspond to the requested increase in resources may be determined. As discussed above a requested increase of resources may be a simple request to increase a single component of the computing system, such as CPU quota, or it may represent an increase in an elastic scaling dimension that may require the adjustment and reallocation of more than one component in the computing system, such as, CPU quota, RAM quota, network bandwidth quota, and the like. In some embodiments, an increase of one resource may result in a decrease of another resource.
Process 1300 next proceeds to block 1304, where the impact that the requested resources may have on the project, other projects, and/or the system in general may be determined. In some embodiments, the impact of the requested resources may be static or dynamic. In at least one embodiment, the impact may be automatically determined by rule-based systems, predictive computing techniques, or the like. In another embodiment, the impact may be manually determined by an operator. Embodiments may be arranged to define rules and thresholds that may set limits on the resources that may be allocated to individual projects. Also, rules may be arranged to determine minimum resources that may be made available for other projects in the system. One of ordinary skill in the art will appreciate that there may be many factors that may be available for analysis when determining if the requested increase of resources may cause unacceptable impact on the project, projects, and/or the system in general. For example, a request to increase index storage size to a very large size may cause such a large data set to be created that the project may not have CPU quota, or RAM quota to effectively process the large dataset. How a condition like this is handled depends on how the embodiment is arranged and configured. In some cases, such an increase in project resources may be aborted because it is out of bounds. In other cases, referring to the example above, an embodiment may determine the necessary CPU quota and RAM quota that may be necessary to keep up with the projected increased index storage size.
After the necessary resources for the requested increase of resources has been determined process 1300 continues at decision block 1306, where a determination may be made whether there is global headspace available for the requested increase of resources. In at least one embodiment, this determination may be made by testing the amount of global headspace available. This test may take into account all of the resources available and provisi— J r~~~ iL ~ operator or process that is requesting the increase of resources for the project. For example, an embodiment may require users to supply a pre-authorization when requested resources would increase a user's total usage of system resources beyond a certain level. Or, an embodiment may be configured to require an administrative user to authorize any resource increase request that may increase total resource use beyond a configured level. In any event, an embodiment may test whether the global conditions may be able to absorb the requested increase in resources.
In at least one embodiment, if there is insufficient global headspace available, then the requested increase in the resource may not be executed at the time of the request. In some embodiments, if the global headspace for the requested increase of resources is available, then process 1300 may flow to decision block 1308; otherwise, process 1300 may flow to block 1312. In some other embodiments, if the increase in resources would exceed global thresholds and/or limitations, then the request for the increase of resources may be discarded.
Process 1300 continues at decision block 1308, where a determination may be made whether the local resources related to the project may be capable of handling the requested increase in project resources. For example, in an embodiment that may be employing virtual machines, the requested increase in project resources may require that a virtual machine allocate more virtual disk space, or it may require that the virtual machine be migrated to another physical machine that may be able to accommodate the requested increase in project resources. In some embodiments, if such operations are unable to be resolved immediately (e.g., within a predefined time period), then the local resources may not be capable of handling the requested increase of project resources. If the local resources are capable of handling the requested increase in project resources, then process 1300 may flow to decision block 1310; otherwise, process 1300 may flow to block 1312 to queue the request.
At decision block 1310, a determination may be made whether the requested increase in project resources may be arranged or configured to occur at a scheduled time in the future. In some embodiments, a user and/or operator may determine a scheduled time when the requested increase may occur. In other embodiments, the scheduled time may be predetermined, such as five minutes from when the request was received. In other embodiments, the scheduled time may be based on the resources requested. For example, if additional CPU quota is requested, then the scheduled time may be one minute from when the request was received, where a request for additional storage space may be scheduled 10 minutes from when the request w~ ~ ~ ' ~~ Λ If the increase may be scheduled to occur in the future, then process 1300 may flow to block 1312 to queue the request; otherwise, process 1300 may flow to block 1314.
At block 1312, the request for the increase of resources may be submitted to a queue. In at least one embodiment, the request may remain in the queue until conditions allow it to be fulfilled (i.e., available resources). One embodiment of queuing project resource requests is described in more detail below in conjunction with FIGURE 15. In some embodiments, a process or operator requesting an increase in project resources may not have knowledge of the global resource allocation. Queuing the requested increase in project resources may enable short running processes to finish and expire so that their released resources may be made available to a queued request for increasing project resources. After block 1312, process 1300 may return control to a calling process.
If the requested increase in project resources is not queued, then process 1300 may flow to block 1314. At block 1314, the determined resources for the requested increase may be allocated. Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
After block 1314, process 1300 may return control to a calling process.
FIGURE 14 depicts a logical flow diagram generally showing one embodiment of process for responding to a requested decrease in project resources. In some embodiments, process 1400 of FIGURE 14 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1400 or portions of process 1400 of FIGURE 14 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1400 or portions of process 1400 of FIGURE 14 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. In some embodiments, a requested decrease in project resources may require some system or computing resources to decrease while others may be required to be increased depending on how the relevant control dimension has been arranged and configured.
Process 1400 begins, after a start block, at decision block 1402, where a determination may be made whether the requested decrease in project resources may be scheduled to occur in the future. In at least one embodiment, decision block 1402 may employ embodiments of decision block 1310 of FIGURE 13 to determine if a request may be scheduled to occur at a time after the request is received. If the decrease in project resources may be scheduled in the future, then process 1400 may flow to block 1410 to queue the request at least until the scheduled time; otherwise, process 1400 may flow to block 1404.
At block 1404, the impact of the request on the project, other projects, and/or the system as a whole may be determined. In at least one embodiment, block 1404 may employ
embodiments of block 1304 of FIGURE 13 to determine an impact of resources on the project and the system (noting that block 1304 of FIGURE 13 determines based on an increase in resources and block 1404 determines based on a decrease in resources).
Process 1400 next proceeds to decision block 1406, where a determination may be made whether the impact of the requested decrease in project resources is unacceptable. In at least one embodiment, running processes of the current project may be analyzed to determine if they can run to completion under the conditions of the decreased resources. In one non-limiting, non- exhaustive example, a current project may not run to completion if the decrease in resources starves the project of processor time, which may be determined to be unacceptable. If the requested decrease in project resources is acceptable, then process 1400 may flow to block 1408; otherwise, process 1400 may flow to block 1410 to queue the request, which may enable the current processes to finish gracefully. But, in some embodiments a requested decrease in project resource may be arranged and configured so it may not cooperate with the running processes. In such cases, an embodiment may terminate the running process to avoid unpredictable outcomes that may occur if the processes try to continue to run reduced resource environment.
At block 1410, the resource decrease request may be queued. An embodiment of a project resource request queue is shown in FIGURE 15. As discussed above project resource requests may be submitted to a request queue if the request may not be completed at the time it may be initiated. This may occur for many reasons, including lack of local or global resources, lack of authorizations that may be needed, the request may be deliberately schedule to occur in the future, currently running processes may be blocking the request, and the like.
A project resource request queue may be arranged to hold request records that may comprise a data structure that carries the metadata that may contain the information about the resources that may have been determined necessary to execute the request. Such metadata may include instructions on how to allocate resources for all of the components impacted by the resource request. For example, a project resource request may require a CPU quota adjustment and index storage size allocation as part of a single project resource request. In a case like this, a queued resource request may contain metadata and instructions that may be used to executed the CPU quota change and the index storage allocation. In other embodiments the queued request may contain the control delta information from the component that initiated the project request and the underlying components and instructions needed to execute the request may be determined at the time of allocation. If, at decision block 1406, the requested decrease in project resources is acceptable, then process 1400 may flow from decision block 1406 to block 1408. At block 1408, resource allocation for the current project may be decreased. Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
After block 1408, process 1400 may return control to a calling process. FIGURE 15 illustrates a logical flow diagram generally showing one embodiment of process for queuing resource requests. In some embodiments, process 1500 of FIGURE 15 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1500 or portions of process 1500 of FIGURE 15 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1500 or portions of process 1500 of FIGURE 15 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Process 1500 begins, after a start block, at decision block 1502, where a determination may be made whether there are queued requests. In at least one embodiment, this determination may be based on whether the queue is holding any requests. If the queue is holding requests, then process 1500 may flow to block 1504; otherwise, process 1500 may return control to a calling process.
At block 1504, a next request in the queue may be examined. In at least one embodiment, this request examination may include a determination of the resources associated with the next request and a determination of the conditions that need to be met for the next request to be acted upon (e.g., acceptable resources available to accommodate the request). In some embodiments, the conditions may include the conditions that led to the next request being entered into the request queue in the first place. For example, the condition may include a scheduled time. Process 1500 next proceeds to decision block 1506, where a determination may be made whether the conditions to execute the next request are satisfied for the next request. For example, a next request that required prior authorization before proceeding may be tested to see if sufficient authorization has been obtained. Or, global resource conditions may be examined to determine if a next request that may have been queued because of a lack of available global resources now may have sufficient resources available to proceed. For next requests that may have been queued so they may occur at a scheduled time in the future the time may be checked. If the conditions for the next request are met, then process 1500 may flow to block 1508;
otherwise, process 1500 may loop to decision block 1502. If conditions for the next request are not met, the process may leave the request in the queue and may examine the next request in the queue. In some embodiments a queue may be configured such that if a request remains in the queue longer that a specified time an additional action may be taken. Such actions may include, notifying the operator, generating alerts in the form of log messages or other notifications, removing the request from the queue, and the like.
At block 1508, the determined resources based on the next request may be allocated. Embodiments for allocating resources are discussed in more detail below in conjunction with FIGURES 16-18.
Process 1500 continues at block 1510, where the completed request may be removed from the queue. In some embodiments, the next request may be configured to be returned to the queue after it has executed. Some requests may be scheduled to occur at a certain time each day. These requests may be returned to the queue to wait until the next time to execute. Other requests may be considered "standing" requests that may remain in the queue waiting for certain conditions to be met. For example, an embodiment may be arranged to automatically increase index storage as it is needed. In this kind of case, an embodiment may use a queued request that may execute when it determines that more index storage is needed. Embodiments may be arranged to have multiple project resource request queues designated for different types of requests. Also, embodiments may employ general purpose queues that may be used for queue project resource requests as well as other types of requests and events. After block 1510, process 1500 may return control to a calling process.
As discussed briefly above, embodiments may be arranged to support elastic scaling in one or more dimensions. FIGURE 16 depicts a logical flow diagram generally showing one embodiment of process for elastically scaling based on index storage size. In some
embodiments, process 1600 of FIGURE 16 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1600 or portions of process 1600 of FIGURE 16 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
embodiments, process 1600 or portions of process 1600 of FIGURE 16 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B Process 1600 begins, after a start block, at block 1602, where an optimal storage size for a new index store may be determined. In some embodiments, optimal storage size may be static or dynamic. In at least one embodiment, optimal storage size may be automatically determined by rule-based systems, predictive computing techniques, or the like. In another embodiment, optimal storage size may be manually determined by an operator. The optimal size may be determined by a rule-based system tailored to the specific project that owns the index store. An embodiment of a rule-based system for elastic scaling may employ triggers linked to performance and resource thresholds. Different projects may be configured to have different triggers and different thresholds depending on the requirements of the project. In addition to rule-based determination, the optimal size for index store may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator has used a control to determine the optimal size, an embodiment may translate the operation of the control into the actual optimal index store size. For example, an operator may use a user interface to control to increase an index store from 5 gigabytes to 10 gigabytes. In this case an embodiment may have to make several determinations to account for the actual storage size needed to account for administration overhead that may not be apparent to the operator.
Process 1600 continues at block 1604, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the index store. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
Process 1600 proceeds next to block 1606, where new objects and/or processes may be allocated and/or instantiated to optimally service the scaled new index store. Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
Continuing at block 1608, the new index store may be determined. At least one embodiment for determining a new index store is described in more detail below in conjunction with FIGURE 19. Briefly, however, in some embodiments indexers may be activated (i.e., marked as read/write) or deactivated (i.e., marked as read-only based on a number of indexers needed for the scaled new index store storage. In some other embodiments, the current index store may be marked as read-only and the new index store may be generated from new indexers. This may ensure that the all new writes to the index store may be written into the scaled new index store storage.
Process 1600 next proceeds to block 1610, which is discussed in more detail below in conjunction with FIGURE 22. Briefly, however, at block 1610, an input proxy may be directed to use the scaled new index store for the indexing of incoming events.
Process 1600 continues at block 1612, where the new objects and/or processes may be employed to index incoming events using the scaled new index store. In at least one
embodiment, incoming events and/or incoming data may include new events/data and/or archived events/data. In at least one embodiment, the scaled new index store may be utilized based on Roles of the new indexes. In some embodiments, new indexes (which may include the new objects, new processes, or the like) may be automatically added to existing indexes (which may include the existing objects, existing processes, or the like). In at least one embodiment, the new indexes may be attached to and/or inherit Roles associated with the existing indexes for at least one Core User of the project. In such an embodiment, the input proxy may utilize the inherited Roles during a scaling event to determine when to send incoming events (and/or other data) to the new indexes. As described above, the Roles may determine access controls to the indexes.
After block 1612, process 1600 may return control to a calling process.
FIGURE 17 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on indexing performance. In some embodiments, process 1700 of FIGURE 17 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1700 or portions of process 1700 of FIGURE 17 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1700 or portions of process 1700 of FIGURE 17 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Indexing performance may be a compound elastic scaling dimension that may be generally related to speed of the processing for a project. For many embodiments multiple component resources, such CPU quota, network bandwidth quota, RAM quota, index storage, and like, may be scaled to execute elastic scaling based on Indexing Performance.
Process 1700 begins, after a start block, at block 1702, where resources needed for optimal indexing performance may be determined. In at least one embodiment, optimal indexing performance may be automatically determined by rule-based systems, predictive computing techniques, or the like. In another embodiment, optimal indexing performance may be manually determined by an operator.
The specific resources needed may be determined by a rule-based system tailored to a specific project. An embodiment of a rule -based system for elastic scaling may employ triggers linked to performance and resource thresholds. Different projects may be configured to have different triggers and different thresholds depending on the requirements of the project. In addition to rule-based determination the optimal resource allocation may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator may use a control to determine the optimal indexing performance an embodiment may translate the operation of the control into the actual optimal allocation of resources. For example, an operator may use a user interface to control to increase a projects indexing performance from "low" to "high." However, in this case an embodiment may have to make several determinations to account for CPU quota, disk access quota, RAM quota, storage size, and the like, the details of which may not be apparent to the operator.
Process 1700 proceeds to block 1704, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the indexing performance. An embodiment may allocate and format a new storage even if the size of the data store has not increase or decrease. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
Processing continues at block 1706, where new objects and/or processes may be allocated and/or instantiated to optimally support the determined indexing performance requirements. Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
Process 1700 proceeds next at block 1708, where the new index store may be
determined. In at least one embodiment, block 1708 may employ embodiments of block 1608 of FGURE 16 to determine the new index store.
Process 1700 continues next at block 1710, which is discussed in more detail below in conjunction with FIGURE 22 Briefly, however, at block 1710, an input proxy may be directed to use the scaled new index store for the indexing of incoming events.
Process 1700 proceeds to block 1712, where the new objects and/or processes may be employed to index incoming events using the scaled new index store. In at least one
embodiment, incoming events and/or incoming data may include new events/data and/or archived events/data. In at least one embodiment, block 1712 may employ embodiments of block 1612 of FIGURE 16 to use the scaled new index store.
After block 1712, process 1700 may return control to a calling process.
FIGURE 18 depicts a logical flow diagram generally showing one embodiment of a process for elastically scaling based on an event time window. In some embodiments, process 1800 of FIGURE 18 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1800 or portions of process 1800 of FIGURE 18 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 1800 or portions of process 1800 of FIGURE 18 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Event time window may be a compound elastic scaling dimension that may be generally related to the index storage sizes for a project. An embodiment may be arranged to receive data such that the index store may be sufficient to hold data received during a determined time window. A time window may be described as date-time expression such as, a start date -time and an end date-time, a duration, such as 30 days, start date-time plus a duration, and the like. For many embodiments multiple component resources, such CPU quota, network bandwidth quota, RAM quota, index storage, and like, may be scaled to execute elastic scaling based on event time windows.
Process 1800 begins, after a start block, at block 1802, where a new event time window may be determined. The event time window may be submitted as a complete time duration expression or it may be dynamically determined based on past event traffic and predicted future event traffic or any method to describe a time window for capturing events.
Process 1800 proceeds next to block 1804, where resources needed for indexing data for the determined event time window may be determined. The specific resources needed may be determined by a rule-based system tailored for specific projects. An embodiment of a rule-based system for elastic scaling may employ triggers linked to performance and resource thresholds. Different projects may be configured to have different triggers and different thresholds depending on the requirements of the project. In addition to rule-based determination, the optimal resource allocation may be determined by an operator using a user-interface control such as a slider, dial, radio buttons, and the like. Even though an operator may use a control to determine the optimal indexing performance an embodiment may translate the operation of the control into the actual optimal allocation of resources. For example, an operator may use a user interface control to increase a project's event time window from "10 days" to "30 days." However, in this case an embodiment may have to make several determinations to account for CPU quota, disk access quota, RAM quota, storage size, and the like, the details of which may not be apparent to the operator. In any event, process 1800 continues at block 1806, where new storage may be allocated and/or formatted to accommodate the elastic scaling of the determined event time window in the new index store. New storage may be created so the current index store may not be altered or disturbed during the elastic scaling process.
Process 1800 proceeds to block 1808, where new objects and/or processes may be allocated and/or instantiated to optimally support the determined event time window.
Embodiments may create new objects and processes to avoid disturbing the current index store during the elastic scaling process.
Continuing to block 1810, the new index store may be determined. In at least one embodiment, block 1708 may employ embodiments of block 1608 of FGURE 16 to determine the new index store.
Process 1800 proceeds next to block 1812, which is discussed in more detail below in conjunction with FIGURE 22. Briefly, however, at block 1812, an input proxy may be directed to use the scaled new index store for the indexing of incoming data.
Process 1800 continues to block 1814, where the new objects and/or processes may be employed for indexing incoming data using the scaled new index store. In at least one embodiment, incoming events and/or incoming data may include new events/data and/or archived events/data. In at least one embodiment, block 1814 may employ embodiments of block 1612 of FIGURE 16 to use the scaled new index store.
After block 1814, process 1800 may return control to a calling process.
FIGURE 19, illustrates a logical flow diagram generally showing one embodiment of a process for changing a number of indexers associated with an index store. In some
embodiments, process 1900 of FIGURE 19 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 1900 or portions of process 1900 of FIGURE 19 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other
embodiments, process 1900 or portions of process 1900 of FIGURE 19 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. Although process 1900 describes changing a number of indexers, in some embodiments, changing a number of indexers (e.g., indexer servers) may also include changing an amount of resources associated with a project, such as, for example, storage, objects, processors, or the like. In at least one embodiment, adding additional indexers may include adding additional storage space and/or processors (and/or other resources). In another embodiment, reducing a number of indexers may include reducing storage space and/or processors (and/or other resources. In yet other embodiments, resources for a given indexer server may also be changed. In some embodiments, resources of an indexer server may be changed by changing a number of indexes used by a project, which may also change an amount of storage, processing capacity, or the like for the project.
Process 1900 begins, after a start block, at decision block 1902, where a determination may be made whether additional indexers may be added to the index store. In at least one embodiment, the determination may be based on the elastic scaling. In some embodiments, if the storage size and/or the performance are increased, then indexers may be added to the index store. If indexers may be added to the index store, then process 1900 may flow to decision block 1904; otherwise, process 1900 may flow to decision block 1906.
At decision block 1906, a determination may be made whether to remove indexers from the index store. In at least one embodiment, the determination may be based on the elastic scaling. In some embodiments, if the storage size and/or the performance are decreased, then indexers may be removed from the index store. If indexers may be removed from the index store, then process 1900 may flow to block 1908; otherwise, process 1900 may return control to a calling process.
At block 1908, at least one indexer may be deactivated. In some embodiments, deactivating an indexer may include marking the indexer as read-only. In at least one embodiment, a read-only indexer may indicate that the indexer is not currently active for the new index store. In some embodiments, a determination may be made on which indexers to deactivate based on time (e.g., creation time, last modified time, last access time, or the like), size of the indexer, or the like. In at least one embodiment, a user may be prevented from accessing a deactivated indexer. In some embodiments, if the indexer is deactivated for a period of time, then the data may be removed from the deactivated indexer (e.g., deleted and/or moved to archival storage). If, at decision block 1902, additional indexers may be added to the index store, then processing may flow from decision block 1902 to decision block 1904. At decision block 1904, a determination may be made whether there are deactivated indexers (such as were deactivated at block 1908). If there are deactivated indexers, then process 1900 may flow to block 1912; otherwise, process 1900 may flow to block 1910.
At block 1912, at least one currently deactivated indexer may be activated and added to the new index store. In at least one embodiment, activating may include marking a read-only indexer as a read/write indexer. After block 1912, process 1900 may return control to a calling process. If, at decision block 1904, there are no deactivated indexers, then process 1900 may flow from decision block 1904 to block 1910. At block 1910, new indexers may be added to the new index store. After block 1910, process 1900 may return control to a calling process.
FIGURE 20 shows a logical flow diagram generally showing one embodiment of a process to recall archived data for re -indexing. In some embodiments, process 2000 of FIGURE 20 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 2000 or portions of process 2000 of FIGURE 20 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 2000 or portions of process 2000 of FIGURE 20 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B.
Embodiments may be arranged to archive input data as it is indexed. (See, input proxy, as shown in FIGURE 22, for more details about archiving.) Data stored in the archives may be in the same format and form as it was sent to the input proxy. A process may process data from the archive and create indexes from the data. This may enable operators to re-create indexes from the archived data and it may enable operators to create new indexes that may be processed differently from the indexes that may have been created when the data was first received.
Process 2000 begins, after a start block, at block 2002, where a time window of events may be determined for re-indexing. The time windows may be determined be a rule-based algorithm that may automatically determine the time window or an operator may select the time window manually using a user interface.
Process 2000 continues at block 2004, where event data records may be retrieved from the archive. Process 2000 proceeds next to block 2000, where the retrieved event data records are indexed using the current project's parameters. Recall of archived event data may occur in the context of a new project or it may occur as part of an existing project.
Processing continues at decision block 2008, where a determination may be made whether, there are additional records remaining in the archive. In some embodiments, the process may continue to retrieve records and index them from the archive until all the data within the determined time window has been exhausted. In some embodiments a recall process may be aborted or paused based on operator intervention or when determined performance thresholds have been reached. If there are records remaining in the archive, then process 2000 may loop to block 2004; otherwise, process 2000 may return control to a calling process. As described above, a new scaled index store may be employed based on elastic scaling.
In some embodiments, data from a previous index store may remain available while the new scaled index store is employed. FIGURE 21 illustrates a logical flow diagram generally showing one embodiment of a process for transitioning to scaled index stores. In some embodiments, process 2100 of FIGURE 21 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 2100 or portions of process 2100 of FIGURE 21 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 2100 or portions of process 2100 of FIGURE 21 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. Process 2100 begins, after a start block, at block 2102, where event data may be received from an input proxy. Embodiments of employing an input proxy are described in more detail below in conjunction with FIGURE 22. Process 2100 next proceeds to block 2104, where the received event data may be indexed. In at least one embodiment, the indexed event data may be stored in the new scaled index store.
Process 2100 continues at block 2106, where the total storage for the new events stored in the scaled new index store and the old event that may be stored in the previously used readonly index store may be determined. Proceeding to decision block 2108, a determination may be made whether the total amount of storage used is greater than the amount of stated provisioned for the project. If the total storage used is greater than the amount of storage provisioned for the project, then process 2100 may flow to block 2110; otherwise, process 2100 may loop back to block 2102 to receive additional event data from the Input Proxy.
At block 2110, the index records in the previously provisioned storage may be marked for archival or removal. For the index records marked for archival, process 2100 may flow to block 2112; for the index records marked for removal, process 2100 may flow to block 2114.
At block 2110, the index records marked for archival may be copied from the previously provisioned storage to archival storage.
At block 2114, the index records marked for removal may be deleted from the previously provisioned storage. In some embodiments, index records marked from archive may be deleted from the previously provisioned storage after being copied to the archival storage.
After block 2114, process 2100 may return control to a calling process.
As described above, some embodiments may enable the operation of an input proxy. FIGURE 22 illustrates a logical flow diagram generally showing one embodiment of an input proxy operation. In some embodiments, process 2200 of FIGURE 22 may be implemented by and/or executed on a single network device, such as network device 400 of FIGURE 4. In other embodiments, process 2200 or portions of process 2200 of FIGURE 22 may be implemented by and/or executed on a plurality of network devices, such as network device 400 of FIGURE 4. In yet other embodiments, process 2200 or portions of process 2200 of FIGURE 22 may be implemented by and/or executed on one or more blade servers, such as blade server 250 of FIGURE 2B. Process 2200 begins, after a start block, at block 2202 where the proxy may receive data. The data may be received from remote clients and sources. The data source may stream the data to the input proxy using various networking technologies, including, Ethernet, 802.11 (a,b,g, and n), Bluetooth, TCP/IP, UDP, HTTP, internal specific point-to-point protocol, and the like. In some embodiments, the proxy may throttle (increase or decrease) a rate at which data incoming data is streamed. In at least one embodiment, the data rate may be based on a corresponding project. For example, the data rate may be based on an amount of storage space allocated for the project (e.g., an amount of storage space purchased by a user for the project). The data may be received and a process may "tee" the data, sending it down at least two paths at the same time, which may be depicted in FIGURE 22 as Path A and Path B.
Proceeding on Path A, process 2200 may proceed from block 2202 to block 2204. At block 2204, the raw data may be load balanced to determine which indexer to use.
Process 2200 continues next at block 2206, where the received data may be indexed on the load balanced indexers. In some embodiments, a size of each indexer may not be limited. In other embodiments, the size of each indexer may be limited. In one such embodiment, an amount of data per indexer may be monitored. In one embodiment, the monitoring may be asynchronous to incoming data. In some embodiment, if a limit for an indexer is reached or almost reached (e.g., less than 5% remaining), then a message may be sent to a user to indicate that the indexer limit is reached or almost reached. In some other embodiments, indexing may be disabled if a user exceeds purchased resource allocation. In at least one embodiment, indexing may be disabled if a user and/or project exceeds an allocated storage size (e.g., previously purchased amount of storage). In another embodiment, indexing may be disabled if a user discontinues paying for the indexing. In yet another embodiment, indexing may be disabled if a user and/or project exceeds a number of buckets for the project.
Process 2200 next proceeds to block 2208, where the indexed data and event records may be stored in the project's index store.
Proceeding on Path B, process 2200 may proceed from block 2202 to block 2210. At block 2210, the raw data may be sent to be copied to intermediate storage. The intermediate storage may be arranged to be available, fast and reliable. Embodiments may employ a FIFO buffer implemented by local hard disks, RAID arrays, and the like, to ensure that the raw data is not lost.
Process 2200 continues at block 2212, where the data may be archived as archival storage becomes available. Moving the raw data from the intermediate storage to the long term archive may be executed by a separate process that reliably moves the raw data from the intermediate storage to the long term storage. After block 2208 and 2212, process 2200 may return control to a calling process.
It will be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by computer program instructions.
These program instructions may be provided to a processor to produce a machine, such that the instructions, which execute on the processor, create means for implementing the actions specified in the flowchart block or blocks. The computer program instructions may be executed by a processor to cause a series of operational steps to be performed by the processor to produce a computer-implemented process such that the instructions, which execute on the processor to provide steps for implementing the actions specified in the flowchart block or blocks. The computer program instructions may also cause at least some of the operational steps shown in the blocks of the flowchart to be performed in parallel. Moreover, some of the steps may also be performed across more than one processor, such as might arise in a multi-processor computer system. In addition, one or more blocks or combinations of blocks in the flowchart illustration may also be performed concurrently with other blocks or combinations of blocks, or even in a different sequence than illustrated.
Accordingly, blocks of the flowchart illustration support combinations of means for performing the specified actions, combinations of steps for performing the specified actions and program instruction means for performing the specified actions. It will also be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by special purpose hardware-based systems, which perform the specified actions or steps, or combinations of special purpose hardware and computer instructions.
Use Case Illustration FIGURE 23 illustrates a non-exhaustive example of a use case of an embodiment of user-interface controls that may be employed to enable elastic scaling. As shown in FIGURE 23, an embodiment may be arranged to provide user-interface controls, such as sliders, dials, radio buttons, spinners, touch controls, gesture controls, and the like, that may enable an operator to elastically scale resources that may be associated with a project. One exemplary user-interface is shown in the FIGURE 23. This may represent a user-interface having three slider controls for controlling the resource for the project.
As shown in FIGURE 23, a slider control (e.g., control 2302, control 2304, and/or control 2306) may generally comprise a slider 2308 and a slider thumb 2310. An operator may manipulate the slider thumb 2310 by moving it forward or backwards lengthwise within the slider 2308. The movement of the slider thumb 2310 may generate intermediate values that may be used for determining corresponding to elastic scaling values.
The embodiment depicted in FIGURE 23 has three slider controls. Control 2302, on the top of the figure, may be use for the elastic scaling of index storage size. Control, in the center of the figure, may be used for the elastic scaling of the number of server instances the may be used to support a project. And, control 2306, at the bottom of the figure, may be used to set the index storage size by event time window.
Control 2302 may also include text box 2312 and text box 2314. In at least one embodiment, text box 2312 may display the current value represented by the position of slider thumb 2310 in slider 2308. In some other embodiments, text box 2312 may enable a user to enter a desired value, and slider thumb 2310 may automatically slide to the corresponding position in slider 2308.
In another embodiment, text box 2314 may display how much the current value (i.e., what is displayed in text box 2312) of control 2302 will cost the owner of the project. In some other embodiments, text box 2314 may enable a user to enter a desired cost, and slider thumb
2310 may automatically slide to the corresponding position in slider 2308. As shown, each slider control may include text box 2312 and/or text box 2314. However, it should be appreciated that embodiments are not restricted to arranging text boxes as depicted herein. Embodiments may be expected to use a variety of user interface methods to indicate the current value of the controls and well as how much the selected value may cost. Also, embodiments may be arranged to display values different and in addition to those depicted herein, or none at all.
Some embodiments may arrange for the user interface controls to use different scales for the variables different that may be represented by the control, such as, linear (as illustrated by control 2302), logarithmic (as illustrated by control 2306), parametric (as illustrated by control 2304), and the like. An embodiment may be arranged to use a particular scale based on the type of user-interface being used and the resource that may be represented. In some cases, the scale of the control may be determined based on how the user may be charged for the underlying resource. For example, if a linear increase in the number of server instances many have a logarithmic increase in the cost to the user it may be advantageous to scale the control to account for the change in cost rather than the change in resources.
In some embodiments multiple controls may be linked together such that adjusting one control may cause another control to adjust as well. Also, embodiments may provide additional user controls (not shown), such as, checkboxes, that may enable a user to lock a control so it may remain in a fixed position as other controls are adjusted.
In other embodiments, some controls may represent high level, generalized, or aggregated representations of resources, in this case a user may select a higher level control and "drill-down" into the underlying elements that the high level, generalized, or aggregated control may be representing.
One of ordinary skill in the art will appreciate that the three controls depicted in the embodiment shown in FIGURE 23 are merely exemplary and many other configurations and arrangements could be used.
The above specification, examples, and data provide a complete description of the composition, manufacture, and use of the invention. Since many embodiments of the invention can be made without departing from the spirit and scope of the invention, the invention resides in the claims hereinafter appended.

Claims

CLAIMS What is claimed is:
1. A method for managing data over a network, wherein at least one network device performs actions, comprising:
initializing at least one project for at least one frontend user, wherein the at least one frontend user is enabled to interact with the at least one project through at least one application;
establishing a data volume limit for an index store that corresponds to the at least one project, wherein the data volume limit is based on a selection by the at least one frontend user;
indexing data that is provided for the at least one project;
if an amount of indexed data in the index store is less than the data volume limit, storing the indexed data in the index store, wherein indexed data in the index store is accessible to the at least one frontend user;
if the amount of indexed data in the index store at least equals the data volume limit, storing the indexed data in an overflow index store that is inaccessible to the at least one frontend user; and
providing at least a portion of indexed data in the overflow index store to the index store if a corresponding amount is added to the data volume limit based on another selection by the frontend user, wherein the provided indexed data is accessible in the index store to the at least one frontend user.
2. The method of Claim 1 , further comprising:
determining a core user that corresponds to a frontend user and a project, wherein the core user enables the frontend user to interact with the project.
3. The method of Claim 1 , further comprising:
copying at least a portion of the indexed data in the overflow index store to archival storage.
4. The method of Claim 1 , wherein the indexed data that exceeds the data volume limit is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit.
5. The method of Claim 1 , wherein providing the indexed data from the overflow index store to the index store further comprises determining an amount of indexed data to copy from the overflow index store to the index store based on at least an increase to the data volume limit.
6. The method of Claim 1 , further comprising:
determining at least one role for each frontend user that establishes permissions for interacting with the at least one project and the at least one application for each frontend user.
7. A network device for managing data over a network, comprising:
a memory for storing data and instructions; and
a processor that executes the instructions to enable actions, including:
initializing at least one project for at least one frontend user, wherein the at least one frontend user is enabled to interact with the at least one project through at least one application;
establishing a data volume limit for an index store that corresponds to the at least one project, wherein the data volume limit is based on a selection by the at least one frontend user;
indexing data that is provided for the at least one project;
if an amount of indexed data in the index store is less than the data volume limit, storing the indexed data in the index store, wherein indexed data in the index store is accessible to the at least one frontend user;
if the amount of indexed data in the index store at least equals the data volume limit, storing the indexed data in an overflow index store that is inaccessible to the at least one frontend user; and
providing at least a portion of indexed data in the overflow index store to the index store if a corresponding amount is added to the data volume limit based on another selection by the frontend user, wherein the provided indexed data is accessible in the index store to the at least one frontend user.
8. The network device of Claim 7, wherein the processer executes further instructions to enable actions, including: determining a core user that corresponds to a frontend user and a project, wherein the core user enables the frontend user to interact with the project.
9. The network device of Claim 7, wherein the processer executes further instructions to enable actions, including:
copying at least a portion of the indexed data in the overflow index store to archival storage.
10. The network device of Claim 7, wherein the indexed data that exceeds the data volume limit is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit.
11. The network device of Claim 7, wherein providing the indexed data from the overflow index store to the index store further comprises determining an amount of indexed data to copy from the overflow index store to the index store based on at least an increase to the data volume limit.
12. The network device of Claim 7, wherein the processer executes further instructions to enable actions, including:
determining at least one role for each frontend user that establishes permissions for interacting with the at least one project and the at least one application for each frontend user.
13. A system for managing data over a network, comprising:
at least one network device, including:
a memory for storing data and instructions; and
a processor that executes the instructions to enable actions, including:
initializing at least one project for at least one frontend user, wherein the at least one frontend user is enabled to interact with the at least one project through at least one application;
establishing a data volume limit for an index store that corresponds to the at least one project, wherein the data volume limit is based on a selection by the at least one frontend user; indexing data that is provided for the at least one project;
if an amount of indexed data in the index store is less than the data volume limit, storing the indexed data in the index store, wherein indexed data in the index store is accessible to the at least one frontend user;
if the amount of indexed data in the index store at least equals the data volume limit, storing the indexed data in an overflow index store that is inaccessible to the at least one frontend user; and
providing at least a portion of indexed data in the overflow index store to the index store if a corresponding amount is added to the data volume limit based on another selection by the frontend user, wherein the provided indexed data is accessible in the index store to the at least one frontend user.
14. The system of Claim 13, wherein the processer executes further instructions to enable actions, including:
determining a core user that corresponds to a frontend user and a project, wherein the core user enables the frontend user to interact with the project.
15. The system of Claim 13, wherein the processer executes further instructions to enable actions, including:
copying at least a portion of the indexed data in the overflow index store to archival storage.
16. The system of Claim 13, wherein the indexed data that exceeds the data volume limit is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit.
17. The system of Claim 13, wherein providing the indexed data from the overflow index store to the index store further comprises determining an amount of indexed data to copy from the overflow index store to the index store based on at least an increase to the data volume limit.
18. The system of Claim 13, wherein the processer executes further instructions to enable actions, including: determining at least one role for each frontend user that establishes permissions for interacting with the at least one project and the at least one application for each frontend user.
19. A processor readable non-transitory storage media that includes instructions for managing data over a network, wherein the execution of the instructions by a processor enables actions, comprising:
initializing at least one project for at least one frontend user, wherein the at least one frontend user is enabled to interact with the at least one project through at least one application;
establishing a data volume limit for an index store that corresponds to the at least one project, wherein the data volume limit is based on a selection by the at least one frontend user;
indexing data that is provided for the at least one project;
if an amount of indexed data in the index store is less than the data volume limit, storing the indexed data in the index store, wherein indexed data in the index store is accessible to the at least one frontend user;
if the amount of indexed data in the index store at least equals the data volume limit, storing the indexed data in an overflow index store that is inaccessible to the at least one frontend user; and
providing at least a portion of indexed data in the overflow index store to the index store if a corresponding amount is added to the data volume limit based on another selection by the frontend user, wherein the provided indexed data is accessible in the index store to the at least one frontend user.
20. The media of Claim 19, further comprising:
determining a core user that corresponds to a frontend user and a project, wherein the core user enables the frontend user to interact with the project.
21. The media of Claim 19, further comprising:
copying at least a portion of the indexed data in the overflow index store to archival storage.
22. The media of Claim 19, wherein the indexed data that exceeds the data volume limit is automatically deleted based on at least one of a first in first out operation of the overflow index store or any indexed data that exceeds the data volume limit.
23. The media of Claim 19, wherein providing the indexed data from the overflow index store to the index store further comprises determining an amount of indexed data to copy from the overflow index store to the index store based on at least an increase to the data volume limit.
24. The media of Claim 19, further comprising:
determining at least one role for each frontend user that establishes permissions for interacting with the at least one project and the at least one application for each frontend user.
PCT/US2012/050452 2011-08-12 2012-08-10 Data volume management WO2013025553A2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201161523063P 2011-08-12 2011-08-12
US61/523,063 2011-08-12
US13/572,434 2012-08-10
US13/572,434 US8682860B2 (en) 2011-08-12 2012-08-10 Data volume management

Publications (2)

Publication Number Publication Date
WO2013025553A2 true WO2013025553A2 (en) 2013-02-21
WO2013025553A3 WO2013025553A3 (en) 2013-04-25

Family

ID=47678182

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/US2012/050456 WO2013025556A1 (en) 2011-08-12 2012-08-10 Elastic scaling of data volume
PCT/US2012/050452 WO2013025553A2 (en) 2011-08-12 2012-08-10 Data volume management

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/US2012/050456 WO2013025556A1 (en) 2011-08-12 2012-08-10 Elastic scaling of data volume

Country Status (2)

Country Link
US (20) US8682860B2 (en)
WO (2) WO2013025556A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8849779B2 (en) 2011-08-12 2014-09-30 Splunk Inc. Elastic scaling of data volume

Families Citing this family (78)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8966085B2 (en) * 2012-01-04 2015-02-24 International Business Machines Corporation Policy-based scaling of computing resources in a networked computing environment
US8595262B1 (en) * 2012-03-29 2013-11-26 Amazon Technologies, Inc. Resource resolution in computing environments using directed graphs
US9742684B1 (en) * 2012-11-01 2017-08-22 Amazon Technologies, Inc. Adaptive service scaling
US20140258316A1 (en) * 2013-03-08 2014-09-11 Open Text S.A. System and Method for Content Assessment
US9524113B2 (en) 2013-05-24 2016-12-20 Seagate Technology Llc Variable redundancy in a solid state drive
CN103401891A (en) * 2013-06-25 2013-11-20 李冰 General wired/wireless network system
US9894022B2 (en) 2013-07-19 2018-02-13 Ambient Consulting, LLC Image with audio conversation system and method
US8583777B1 (en) * 2013-08-13 2013-11-12 Joingo, Llc Method and system for providing real-time end-user WiFi quality data
US10180776B2 (en) 2013-10-01 2019-01-15 Ambient Consulting, LLC Image grouping with audio commentaries system and method
US10057731B2 (en) 2013-10-01 2018-08-21 Ambient Consulting, LLC Image and message integration system and method
US9977591B2 (en) 2013-10-01 2018-05-22 Ambient Consulting, LLC Image with audio conversation system and method
US20150092006A1 (en) * 2013-10-01 2015-04-02 Filmstrip, Inc. Image with audio conversation system and method utilizing a wearable mobile device
US20150149474A1 (en) * 2013-11-27 2015-05-28 KCura Corporation Method and apparatus for indexing and searching documents
US11809451B2 (en) 2014-02-19 2023-11-07 Snowflake Inc. Caching systems and methods
US9693010B2 (en) * 2014-03-11 2017-06-27 Sony Corporation Method, electronic device, and server for generating digitally processed pictures
US9052938B1 (en) * 2014-04-15 2015-06-09 Splunk Inc. Correlation and associated display of virtual machine data and storage performance data
US10454970B2 (en) * 2014-06-30 2019-10-22 Vescel, Llc Authorization of access to a data resource in addition to specific actions to be performed on the data resource based on an authorized context enforced by a use policy
US9959301B2 (en) 2014-07-25 2018-05-01 Cisco Technology, Inc. Distributing and processing streams over one or more networks for on-the-fly schema evolution
US9438676B2 (en) 2014-07-25 2016-09-06 Cisco Technology, Inc. Speculative data processing of streaming data
US10318896B1 (en) * 2014-09-19 2019-06-11 Amazon Technologies, Inc. Computing resource forecasting and optimization
EP3015981B1 (en) * 2014-10-31 2018-07-25 Khalifa University of Science, Technology and Research Networked resource provisioning system
EP3021274A1 (en) * 2014-11-11 2016-05-18 Tata Consultancy Services Limited Data privacy management
US10572863B2 (en) 2015-01-30 2020-02-25 Splunk Inc. Systems and methods for managing allocation of machine data storage
US9922114B2 (en) * 2015-01-30 2018-03-20 Splunk Inc. Systems and methods for distributing indexer configurations
CN106161492B (en) * 2015-03-24 2019-05-10 华为技术有限公司 A kind of decision data method, apparatus and system
CN104778244B (en) * 2015-04-09 2018-08-24 天脉聚源(北京)传媒科技有限公司 The searching method and device of data
US10282455B2 (en) 2015-04-20 2019-05-07 Splunk Inc. Display of data ingestion information based on counting generated events
US10817544B2 (en) * 2015-04-20 2020-10-27 Splunk Inc. Scaling available storage based on counting generated events
US10057187B1 (en) * 2015-05-27 2018-08-21 Amazon Technologies, Inc. Dynamic resource creation to connect client resources in a distributed system
US10454938B2 (en) * 2015-05-28 2019-10-22 International Business Machines Corporation Dynamic permission roles for cloud based applications
CN105162826A (en) * 2015-07-15 2015-12-16 中山大学 Cloud computing multilayer cloud architecture
US10235431B2 (en) * 2016-01-29 2019-03-19 Splunk Inc. Optimizing index file sizes based on indexed data storage conditions
CN107316189B (en) * 2016-04-27 2022-03-08 华为技术有限公司 User resource processing method and device
US11327475B2 (en) 2016-05-09 2022-05-10 Strong Force Iot Portfolio 2016, Llc Methods and systems for intelligent collection and analysis of vehicle data
US11009865B2 (en) 2016-05-09 2021-05-18 Strong Force Iot Portfolio 2016, Llc Methods and systems for a noise pattern data marketplace in an industrial internet of things environment
US11774944B2 (en) 2016-05-09 2023-10-03 Strong Force Iot Portfolio 2016, Llc Methods and systems for the industrial internet of things
US10108328B2 (en) * 2016-05-20 2018-10-23 Vmware, Inc. Method for linking selectable parameters within a graphical user interface
CN107515879B (en) * 2016-06-16 2021-03-19 伊姆西Ip控股有限责任公司 Method and electronic equipment for document retrieval
US10108447B2 (en) 2016-08-30 2018-10-23 Vmware, Inc. Method for connecting a local virtualization infrastructure with a cloud-based virtualization infrastructure
US10157071B2 (en) 2016-08-30 2018-12-18 Vmware, Inc. Method for migrating a virtual machine between a local virtualization infrastructure and a cloud-based virtualization infrastructure
US10216862B1 (en) * 2016-09-26 2019-02-26 Splunk Inc. Predictive estimation for ingestion, performance and utilization in a data indexing and query system
US10462057B1 (en) * 2016-09-28 2019-10-29 Amazon Technologies, Inc. Shaping network traffic using throttling decisions
US9742549B1 (en) * 2016-09-29 2017-08-22 Analog Devices Global Apparatus and methods for asynchronous clock mapping
US10404614B2 (en) * 2016-11-21 2019-09-03 Vmware, Inc. Multi-cloud resource allocation
CN106713337B (en) * 2017-01-03 2020-04-21 北京并行科技股份有限公司 Method and system for accessing super computing center and scheduling server
US10545798B2 (en) * 2017-01-31 2020-01-28 Splunk Inc. Resegmenting chunks of data for efficient load balancing across indexers
US10489720B2 (en) 2017-05-02 2019-11-26 Secureworks Corp. System and method for vendor agnostic automatic supplementary intelligence propagation
CN107204978B (en) * 2017-05-24 2019-10-15 北京邮电大学 A kind of access control method and device based on multi-tenant cloud environment
WO2018237195A1 (en) * 2017-06-21 2018-12-27 Rational Solutions, Llc Precision professional health-related (phr) communication systems and related interfaces
US11131989B2 (en) 2017-08-02 2021-09-28 Strong Force Iot Portfolio 2016, Llc Systems and methods for data collection including pattern recognition
CN107480237B (en) * 2017-08-08 2021-02-05 中南大学 Data fusion method and system for heterogeneous desktop cloud platform
US11100152B2 (en) 2017-08-17 2021-08-24 Target Brands, Inc. Data portal
CN107633051A (en) * 2017-09-15 2018-01-26 努比亚技术有限公司 Desktop searching method, mobile terminal and computer-readable recording medium
US10484490B2 (en) * 2017-10-05 2019-11-19 Bank Of America Corporation Multicomputer processing of user data with centralized event control
US11397726B2 (en) * 2017-11-15 2022-07-26 Sumo Logic, Inc. Data enrichment and augmentation
US11182434B2 (en) 2017-11-15 2021-11-23 Sumo Logic, Inc. Cardinality of time series
US11113097B2 (en) 2017-12-13 2021-09-07 Boomi, Inc. System and method for provisioning integration infrastructure at runtime indifferent to hybrid nature of endpoint applications
US11074272B1 (en) 2017-12-21 2021-07-27 Seeq Corporation System and method for managing streaming calculations
US10936980B2 (en) * 2018-05-07 2021-03-02 Servicenow, Inc. Graphical user interface for enterprise project portfolio management utilizing normalized project attributes mapped to a common framework
CN108875084B (en) * 2018-07-17 2021-03-12 上海精数信息科技有限公司 Project sound volume sorting method and device, electronic equipment and storage medium
KR20210044226A (en) 2018-08-30 2021-04-22 쉘 인터내셔날 리써취 마트샤피지 비.브이. Heavy base oil and manufacturing method without haze at 0℃
CN110971942B (en) 2018-09-28 2021-08-17 杭州海康威视数字技术股份有限公司 Data transmission method and device
JP7131357B2 (en) * 2018-12-12 2022-09-06 富士通株式会社 Communication device, communication method, and communication program
US10521605B1 (en) 2019-03-15 2019-12-31 ZenPayroll, Inc. Tagging and auditing sensitive information in a database environment
US11200338B2 (en) 2019-03-15 2021-12-14 ZenPayroll, Inc. Tagging and auditing sensitive information in a database environment
CN110035590B (en) * 2019-04-22 2021-08-17 深圳仙苗科技有限公司 Low-standby-power-consumption intelligent lamp control system applying novel networking technology
US11361001B2 (en) * 2019-06-27 2022-06-14 Sigma Computing, Inc. Search using data warehouse grants
CN110677321B (en) * 2019-09-25 2021-09-17 广州通导信息技术服务有限公司 Elastically-telescopic cloud pressure measuring method, device, equipment and storage medium
CN110704397A (en) * 2019-09-27 2020-01-17 宁波和利时智能科技有限公司 Data query method and device based on elastic search
US11469943B2 (en) * 2019-12-06 2022-10-11 Red Hat, Inc. Pre-scheduling for cloud resource provisioning
US11438841B2 (en) * 2020-01-31 2022-09-06 Dell Products, Lp Energy savings system based machine learning of wireless performance activity for mobile information handling system connected to plural wireless networks
CN112765666A (en) * 2021-01-29 2021-05-07 树根互联股份有限公司 Service account data pushing system and method
US11658917B2 (en) * 2021-04-09 2023-05-23 Tekion Corp Selective offloading of bandwidth to enable large-scale data indexing
CN113487245B (en) * 2021-09-06 2021-12-07 苏州浪潮智能科技有限公司 Cross-project resource transfer method and system for cloud platform and computer storage medium
US11929891B1 (en) * 2023-01-10 2024-03-12 Dell Products L.P. System and method for distributed management of hardware through relationship management
US11907230B1 (en) 2023-01-10 2024-02-20 Dell Products L.P. System and method for distributed management of hardware based on intent
US11831706B1 (en) 2023-01-10 2023-11-28 Dell Products L.P. System and method for distributed management of storage systems based on intent
US11770456B1 (en) * 2023-01-10 2023-09-26 Dell Products L.P. System and method for distributed management of storage systems based on subscription changes

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6292795B1 (en) * 1998-05-30 2001-09-18 International Business Machines Corporation Indexed file system and a method and a mechanism for accessing data records from such a system
US20050044198A1 (en) * 2003-08-08 2005-02-24 Jun Okitsu Method of controlling total disk usage amount in virtualized and unified network storage system
US20080021865A1 (en) * 2006-07-20 2008-01-24 International Business Machines Corporation Method, system, and computer program product for dynamically determining data placement
US20090089334A1 (en) * 2007-09-27 2009-04-02 Microsoft Corporation Lazy updates to indexes in a database
US7676510B1 (en) * 2006-12-22 2010-03-09 Network Appliance, Inc. Space reservation monitoring in a fractionally reserved data storage system

Family Cites Families (119)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE1602217B2 (en) 1967-09-30 1971-12-30 Moeller & Neumann Gmbh, 6670 St Ingbert ACCELERATION ROLLER WITH A DEVICE FOR THE PERIODIC THROWING OF CONTINUOUSLY ACCELERATING ROLLER BARS
US5381539A (en) * 1992-06-04 1995-01-10 Emc Corporation System and method for dynamically controlling cache management
US5550971A (en) 1993-06-30 1996-08-27 U S West Technologies, Inc. Method and system for generating a user interface adaptable to various database management systems
US5974409A (en) * 1995-08-23 1999-10-26 Microsoft Corporation System and method for locating information in an on-line network
US6415319B1 (en) 1997-02-07 2002-07-02 Sun Microsystems, Inc. Intelligent network browser using incremental conceptual indexer
US5950465A (en) * 1998-05-26 1999-09-14 Schultz; Theodore J. Deadlock bolt security lock
US6396516B1 (en) 1998-05-29 2002-05-28 Plexus Systems, Llc Graphical user interface shop floor control system
US6247097B1 (en) * 1999-01-22 2001-06-12 International Business Machines Corporation Aligned instruction cache handling of instruction fetches across multiple predicted branch instructions
US6385604B1 (en) 1999-08-04 2002-05-07 Hyperroll, Israel Limited Relational database management system having integrated non-relational multi-dimensional data store of aggregated data elements
US6772348B1 (en) 2000-04-27 2004-08-03 Microsoft Corporation Method and system for retrieving security information for secured transmission of network communication streams
US7284244B1 (en) * 2000-05-02 2007-10-16 Microsoft Corporation Resource manager architecture with dynamic resource allocation among multiple configurations
JP3882479B2 (en) * 2000-08-01 2007-02-14 コクヨ株式会社 Project activity support system
CA2326368A1 (en) * 2000-11-20 2002-05-20 Adexact Corporation Method and system for targeted content delivery, presentation, management, and reporting
EP1217543A1 (en) * 2000-12-22 2002-06-26 Hewlett-Packard Company, A Delaware Corporation Process and apparatus for automatically indexing documents of a set of computers of a network
US6701311B2 (en) * 2001-02-07 2004-03-02 International Business Machines Corporation Customer self service system for resource search and selection
US6870926B2 (en) * 2001-11-06 2005-03-22 Rockwell Electronic Commerce Technologies, Llc Method of optimizing call center resources based upon statistics
US7107285B2 (en) * 2002-03-16 2006-09-12 Questerra Corporation Method, system, and program for an improved enterprise spatial system
US6657565B2 (en) * 2002-03-21 2003-12-02 International Business Machines Corporation Method and system for improving lossless compression efficiency
US7016919B2 (en) 2002-03-29 2006-03-21 Agilent Technologies, Inc. Enterprise framework and applications supporting meta-data and data traceability requirements
US20030191770A1 (en) 2002-04-09 2003-10-09 Nudi Christopher S. System and method for formatting, proofreading and annotating information to be printed in directories
US7246114B2 (en) 2002-10-21 2007-07-17 Microsoft Corporation System and method for presenting a query expressed in terms of an object model
US20040107131A1 (en) * 2002-11-15 2004-06-03 Wilkerson Shawn R. Value innovation management system and methods
US7774191B2 (en) * 2003-04-09 2010-08-10 Gary Charles Berkowitz Virtual supercomputer
US7137019B2 (en) * 2003-04-30 2006-11-14 International Business Machines Corporation Adaptive throttling system for data processing systems
US20040220965A1 (en) 2003-04-30 2004-11-04 Michael Harville Indexed database structures and methods for searching path-enhanced multimedia
US7340480B2 (en) 2003-05-08 2008-03-04 International Business Machines Corporation Iterative data analysis enabled through query result abstraction
US7454569B2 (en) * 2003-06-25 2008-11-18 Commvault Systems, Inc. Hierarchical system and method for performing storage operations in a computer network
US7653688B2 (en) * 2003-11-05 2010-01-26 Sap Ag Role-based portal to a workplace system
US7596571B2 (en) * 2004-06-30 2009-09-29 Technorati, Inc. Ecosystem method of aggregation and search and related techniques
US7562069B1 (en) 2004-07-01 2009-07-14 Aol Llc Query disambiguation
US7299171B2 (en) * 2004-08-17 2007-11-20 Contentguard Holdings, Inc. Method and system for processing grammar-based legality expressions
US20060047576A1 (en) * 2004-08-30 2006-03-02 Nokia Corporation Automated memory and accessory purchasing agent for use in a mobile terminal
US20060095690A1 (en) * 2004-10-29 2006-05-04 International Business Machines Corporation System, method, and storage medium for shared key index space for memory regions
US20060143617A1 (en) 2004-12-29 2006-06-29 Knauerhase Robert C Method, apparatus and system for dynamic allocation of virtual platform resources
US7571157B2 (en) 2004-12-29 2009-08-04 Aol Llc Filtering search results
US8713000B1 (en) 2005-01-12 2014-04-29 Linkedin Corporation Method and system for leveraging the power of one's social-network in an online marketplace
US20060253423A1 (en) 2005-05-07 2006-11-09 Mclane Mark Information retrieval system and method
US8024778B2 (en) 2005-05-24 2011-09-20 CRIF Corporation System and method for defining attributes, decision rules, or both, for remote execution, claim set I
US7730523B1 (en) * 2005-06-17 2010-06-01 Oracle America, Inc. Role-based access using combinatorial inheritance and randomized conjugates in an internet hosted environment
US8200676B2 (en) 2005-06-28 2012-06-12 Nokia Corporation User interface for geographic search
US7937344B2 (en) 2005-07-25 2011-05-03 Splunk Inc. Machine data web
US20070043639A1 (en) * 2005-08-19 2007-02-22 Tabs Mark A Systems and methods for monitoring financial positions
JP4534940B2 (en) * 2005-10-06 2010-09-01 ソニー株式会社 Information recording apparatus, imaging apparatus, information recording control method, and computer program
JP2007148545A (en) 2005-11-24 2007-06-14 Brother Ind Ltd Information delivery system, information delivery method, node device and node processing program
US7958509B2 (en) 2005-12-21 2011-06-07 International Business Machines Corporation Method and system for scheduling of jobs
US7657512B2 (en) * 2005-12-29 2010-02-02 Sap Ag Source-context aware object-based navigation
US7593927B2 (en) 2006-03-10 2009-09-22 Microsoft Corporation Unstructured data in a mining model language
US7552099B2 (en) * 2006-03-10 2009-06-23 International Business Machines Corporation System and method for indexing a data stream
JP4838610B2 (en) * 2006-03-24 2011-12-14 キヤノン株式会社 Document management apparatus, document management method, and program
JP2007304794A (en) * 2006-05-10 2007-11-22 Hitachi Ltd Storage system and storage control method in storage system
US20080059177A1 (en) * 2006-05-19 2008-03-06 Jamey Poirier Enhancement of simultaneous multi-user real-time speech recognition system
US9213513B2 (en) * 2006-06-23 2015-12-15 Microsoft Technology Licensing, Llc Maintaining synchronization of virtual machine image differences across server and host computers
US9392078B2 (en) * 2006-06-23 2016-07-12 Microsoft Technology Licensing, Llc Remote network access via virtual machine
US8392565B2 (en) * 2006-07-20 2013-03-05 Oracle America, Inc. Network memory pools for packet destinations and virtual machines
US20080077538A1 (en) 2006-09-26 2008-03-27 Gregory Schebece Public Trade Execution Analytics
FR2906624A1 (en) * 2006-10-03 2008-04-04 Bull S A S Soc Par Actions Sim Generated data storing system, has management module utilizing calculation result for managing priorities in accessing time by system for copying data of new volumes of virtual bookshop to virtual volumes of physical bookshop
EP2482209A1 (en) 2006-10-05 2012-08-01 Splunk Inc. Time series search engine
US20080147487A1 (en) * 2006-10-06 2008-06-19 Technorati Inc. Methods and apparatus for conversational advertising
US20080104542A1 (en) 2006-10-27 2008-05-01 Information Builders, Inc. Apparatus and Method for Conducting Searches with a Search Engine for Unstructured Data to Retrieve Records Enriched with Structured Data and Generate Reports Based Thereon
US8655939B2 (en) * 2007-01-05 2014-02-18 Digital Doors, Inc. Electromagnetic pulse (EMP) hardened information infrastructure with extractor, cloud dispersal, secure storage, content analysis and classification and method therefor
US8468244B2 (en) * 2007-01-05 2013-06-18 Digital Doors, Inc. Digital information infrastructure and method for security designated data and with granular data stores
US8010991B2 (en) * 2007-01-29 2011-08-30 Cisco Technology, Inc. Policy resolution in an entitlement management system
US20080201118A1 (en) * 2007-02-16 2008-08-21 Fan Luo Modeling a web page on top of HTML elements level by encapsulating the details of HTML elements in a component, building a web page, a website and website syndication on browser-based user interface
US8631147B2 (en) 2007-03-12 2014-01-14 Citrix Systems, Inc. Systems and methods for configuring policy bank invocations
WO2008132924A1 (en) * 2007-04-13 2008-11-06 Nec Corporation Virtual computer system and its optimization method
DE102007043657B4 (en) * 2007-09-13 2009-06-18 Siemens Ag Satellite-overlapping storage organization for medical image data
US20090076887A1 (en) * 2007-09-16 2009-03-19 Nova Spivack System And Method Of Collecting Market-Related Data Via A Web-Based Networking Environment
US8396838B2 (en) * 2007-10-17 2013-03-12 Commvault Systems, Inc. Legal compliance, electronic discovery and electronic document handling of online and offline copies of data
JP2011512126A (en) 2008-02-04 2011-04-21 ザ・コカ−コーラ・カンパニー How to make a custom beverage product
US20090265747A1 (en) * 2008-03-17 2009-10-22 Canada Anv Systems Inc. Systems and methods for providing web based self serviced video monitoring and security features for systems comprising ip video terminals and servers
US8117464B1 (en) * 2008-04-30 2012-02-14 Netapp, Inc. Sub-volume level security for deduplicated data
US20090300065A1 (en) 2008-05-30 2009-12-03 Birchall James T Computer system and methods for improving identification of subrogation opportunities
US7496919B1 (en) * 2008-06-04 2009-02-24 International Business Machines Corporation Method to support role based prioritization of processes
US7987192B2 (en) 2008-06-25 2011-07-26 International Business Machines Corporation Hybrid data model and user interaction for data sets in a user interface
ES2334316B1 (en) * 2008-08-13 2011-06-13 Starlab Barcelona, S.L. METHOD AND SYSTEM OF MULTISITIAN CRANIAL STIMULATION.
US8230510B1 (en) * 2008-10-02 2012-07-24 Trend Micro Incorporated Scanning computer data for malicious codes using a remote server computer
US20100251100A1 (en) 2009-03-31 2010-09-30 Yann Delacourt Primitive-based presentation of dimensional information using layout script
US8548997B1 (en) * 2009-04-08 2013-10-01 Jianqing Wu Discovery information management system
US8850041B2 (en) * 2009-05-26 2014-09-30 Microsoft Corporation Role based delegated administration model
US8606827B2 (en) 2009-05-29 2013-12-10 Red Hat, Inc. Systems and methods for extracting database dimensions as data modeling object
US20120147192A1 (en) * 2009-09-01 2012-06-14 Demaher Industrial Cameras Pty Limited Video camera system
US8935702B2 (en) * 2009-09-04 2015-01-13 International Business Machines Corporation Resource optimization for parallel data integration
US8311964B1 (en) * 2009-11-12 2012-11-13 Symantec Corporation Progressive sampling for deduplication indexing
JP4838878B2 (en) * 2009-12-04 2011-12-14 富士通株式会社 Data management program, data management apparatus, and data management method
US20110137884A1 (en) * 2009-12-09 2011-06-09 Anantharajan Sathyakhala Techniques for automatically integrating search features within an application
US8606667B2 (en) * 2010-02-26 2013-12-10 Red Hat, Inc. Systems and methods for managing a software subscription in a cloud network
CN102236762A (en) * 2010-04-30 2011-11-09 国际商业机器公司 Method for processing file access for multi-tenancy application and file agent device
US8788590B2 (en) * 2010-04-30 2014-07-22 Iliv Technologies Inc. Collaboration tool
US8352387B2 (en) * 2010-05-20 2013-01-08 Verizon Patent And Licensing Inc. Legacy system support
US9135585B2 (en) * 2010-06-15 2015-09-15 Sap Se Managing consistent interfaces for property library, property list template, quantity conversion virtual object, and supplier property specification business objects across heterogeneous systems
AU2010202740B1 (en) * 2010-06-30 2010-12-23 Brightcove Inc. Dynamic indexing for ad insertion in media streaming
AU2010202741B1 (en) * 2010-06-30 2010-12-23 Brightcove Inc. Dynamic chunking for media streaming
US20120030355A1 (en) * 2010-07-27 2012-02-02 Microsoft Corporation Dynamically allocating index server resources to partner entities
US8495625B1 (en) * 2010-07-27 2013-07-23 Symantec Corporation Method and system for creation of streamed files on-demand
US9413677B1 (en) * 2010-09-17 2016-08-09 Sprint Spectrum L.P. Dynamic adjustment of reverse-link rate-control parameters
US20120096497A1 (en) * 2010-10-14 2012-04-19 Sony Corporation Recording television content
US8442982B2 (en) 2010-11-05 2013-05-14 Apple Inc. Extended database search
US20120143928A1 (en) * 2010-12-07 2012-06-07 Filip Yankov File system transaction log flush optimization
US20130290494A1 (en) * 2010-12-22 2013-10-31 Rambus Inc. Session management for communication in a heterogeneous network
US8620897B2 (en) * 2011-03-11 2013-12-31 Microsoft Corporation Indexing and searching features including using reusable index fields
US20120239681A1 (en) 2011-03-14 2012-09-20 Splunk Inc. Scalable interactive display of distributed data
US20120244510A1 (en) 2011-03-22 2012-09-27 Watkins Jr Robert Todd Normalization and Cumulative Analysis of Cognitive Educational Outcome Elements and Related Interactive Report Summaries
US8869244B1 (en) * 2011-05-03 2014-10-21 Symantec Corporation Techniques for providing role-based access control using dynamic shared accounts
US8660789B2 (en) 2011-05-03 2014-02-25 University Of Southern California Hierarchical and exact fastest path computation in time-dependent spatial networks
US8887047B2 (en) 2011-06-24 2014-11-11 Breakthrough Performancetech, Llc Methods and systems for dynamically generating a training program
US20130041874A1 (en) 2011-08-10 2013-02-14 Susan M. Dohm Interactive documentation system for database parameters
US9355394B2 (en) 2011-08-11 2016-05-31 Visa International Service Association Systems and methods of aggregating split payments using a settlement ecosystem
WO2013025556A1 (en) 2011-08-12 2013-02-21 Splunk Inc. Elastic scaling of data volume
US9569511B2 (en) 2011-08-25 2017-02-14 Salesforce.Com, Inc. Dynamic data management
US20130060912A1 (en) 2011-09-07 2013-03-07 Reality Mobile Llc Streaming-Content Analytics
US8914515B2 (en) * 2011-10-28 2014-12-16 International Business Machines Corporation Cloud optimization using workload analysis
US9799040B2 (en) 2012-03-27 2017-10-24 Iprova Sarl Method and apparatus for computer assisted innovation
US9087306B2 (en) 2012-07-13 2015-07-21 Sas Institute Inc. Computer-implemented systems and methods for time series exploration
US20140046976A1 (en) 2012-08-11 2014-02-13 Guangsheng Zhang Systems, methods, and user interface for effectively presenting information
US8788525B2 (en) 2012-09-07 2014-07-22 Splunk Inc. Data model for machine data for semantic search
US9215240B2 (en) 2013-07-25 2015-12-15 Splunk Inc. Investigative and dynamic detection of potential security-threat indicators from events in big data
US10127258B2 (en) 2014-09-30 2018-11-13 Splunk Inc. Event time selection output techniques
US9208463B1 (en) 2014-10-09 2015-12-08 Splunk Inc. Thresholds for key performance indicators derived from machine data
US10855793B2 (en) 2017-09-25 2020-12-01 Splunk Inc. Proxying hypertext transfer protocol (HTTP) requests for microservices

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6292795B1 (en) * 1998-05-30 2001-09-18 International Business Machines Corporation Indexed file system and a method and a mechanism for accessing data records from such a system
US20050044198A1 (en) * 2003-08-08 2005-02-24 Jun Okitsu Method of controlling total disk usage amount in virtualized and unified network storage system
US20080021865A1 (en) * 2006-07-20 2008-01-24 International Business Machines Corporation Method, system, and computer program product for dynamically determining data placement
US7676510B1 (en) * 2006-12-22 2010-03-09 Network Appliance, Inc. Space reservation monitoring in a fractionally reserved data storage system
US20090089334A1 (en) * 2007-09-27 2009-04-02 Microsoft Corporation Lazy updates to indexes in a database

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8849779B2 (en) 2011-08-12 2014-09-30 Splunk Inc. Elastic scaling of data volume
US9037562B2 (en) 2011-08-12 2015-05-19 Splunk Inc. Elastic scaling of data volume
US9225724B2 (en) 2011-08-12 2015-12-29 Splunk Inc. Elastic resource scaling
US9356934B2 (en) 2011-08-12 2016-05-31 Splunk Inc. Data volume scaling for storing indexed data
US9497199B2 (en) 2011-08-12 2016-11-15 Splunk Inc. Access control for event data stored in cloud-based data stores
US9516029B2 (en) 2011-08-12 2016-12-06 Splunk Inc. Searching indexed data based on user roles
US9871803B2 (en) 2011-08-12 2018-01-16 Splunk Inc. Access control for event data stored in cloud-based data stores based on inherited roles
US9992208B2 (en) 2011-08-12 2018-06-05 Splunk Inc. Role-based application program operations on machine data in a multi-tenant environment
US10362041B2 (en) 2011-08-12 2019-07-23 Splunk Inc. Optimizing resource allocation for projects executing in a cloud-based environment
US10616236B2 (en) 2011-08-12 2020-04-07 Splunk Inc. Enabling role-based operations to be performed on machine data in a machine environment
US10887320B1 (en) 2011-08-12 2021-01-05 Splunk Inc. Optimizing resource allocation for projects executing in a cloud-based environment
US11258803B2 (en) 2011-08-12 2022-02-22 Splunk Inc. Enabling role-based operations to be performed on machine data in a machine environment
US11546343B1 (en) 2011-08-12 2023-01-03 Splunk Inc. Optimizing resource allocation for projects executing in a cloud-based environment
US11831649B1 (en) 2011-08-12 2023-11-28 Splunk Inc. Optimizing resource allocation for projects executing in a cloud-based environment
US11855998B1 (en) 2011-08-12 2023-12-26 Splunk Inc. Enabling role-based operations to be performed on machine data in a machine environment

Also Published As

Publication number Publication date
US20180145989A1 (en) 2018-05-24
US20130054537A1 (en) 2013-02-28
US10616236B2 (en) 2020-04-07
US9225724B2 (en) 2015-12-29
WO2013025556A1 (en) 2013-02-21
US20130054814A1 (en) 2013-02-28
US11258803B2 (en) 2022-02-22
US20170048247A1 (en) 2017-02-16
US9497199B2 (en) 2016-11-15
US20130041871A1 (en) 2013-02-14
WO2013025553A3 (en) 2013-04-25
US20160164873A1 (en) 2016-06-09
US20180248890A1 (en) 2018-08-30
US8849779B2 (en) 2014-09-30
US9055075B2 (en) 2015-06-09
US20150269005A1 (en) 2015-09-24
US20170048255A1 (en) 2017-02-16
US9356934B2 (en) 2016-05-31
US8682930B2 (en) 2014-03-25
US9516029B2 (en) 2016-12-06
US9037562B2 (en) 2015-05-19
US11855998B1 (en) 2023-12-26
US20150220585A1 (en) 2015-08-06
US11831649B1 (en) 2023-11-28
US8682860B2 (en) 2014-03-25
US11546343B1 (en) 2023-01-03
US20140297521A1 (en) 2014-10-02
US9992208B2 (en) 2018-06-05
US20200195655A1 (en) 2020-06-18
US10362041B2 (en) 2019-07-23
US20140059036A1 (en) 2014-02-27
US20130042008A1 (en) 2013-02-14
US20150281243A1 (en) 2015-10-01
US9871803B2 (en) 2018-01-16
US20140344900A1 (en) 2014-11-20
US10887320B1 (en) 2021-01-05

Similar Documents

Publication Publication Date Title
US11831649B1 (en) Optimizing resource allocation for projects executing in a cloud-based environment
US11914562B1 (en) Generating search results based on intermediate summaries
US10579648B2 (en) Determining events associated with a value
US11729269B1 (en) Bandwidth management in distributed file systems
US10614033B1 (en) Client aware pre-fetch policy scoring system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12823544

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12823544

Country of ref document: EP

Kind code of ref document: A2