Ascenium Inc

United States of America

Back to Profile

1-62 of 62 for Ascenium Inc Sort by
Query
Aggregations
Jurisdiction
        United States 35
        World 27
Date
2026 December 1
2025 November 1
2025 10
2024 18
2023 16
See more
IPC Class
G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode 39
G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead 33
G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors 23
G06F 8/41 - Compilation 18
G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt 16
See more
Status
Pending 29
Registered / In Force 33
Found results for  patents

1.

MEMORY DEPENDENCE PREDICTION IN A PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number 19235822
Status Pending
Filing Date 2025-06-12
First Publication Date 2025-12-18
Owner Ascenium, Inc. (USA)
Inventor Fredriksson, Hans Olle Viktor

Abstract

A processing unit is accessed that includes a plurality of compute slices, a control unit, and a global aliasing table (GAT). Each compute slice within the plurality of compute slices includes at least one execution unit, is known to a compiler, and is coupled to a successor compute slice and a predecessor compute slice. A first compute slice executes a load instruction. The load instruction is associated with a target address. The load instruction is predicted that it will alias with a previous store instruction. The previous store instruction executes on a previous compute slice among the plurality of compute slices. The predicting is based on the GAT. The load instruction is stalled until the previous store instruction completes execution on the previous compute slice. The load instruction is allowed to execute. The predicting includes searching, in the GAT, for an entry which includes the load instruction.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/355 - Indexed addressing

2.

GLOBAL MEMORY DISAMBIGUATION FOR A PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number 19197924
Status Pending
Filing Date 2025-05-02
First Publication Date 2025-11-06
Owner Ascenium, Inc. (USA)
Inventor
  • Harboe, Øyvind
  • Taylor, Jacob John Vorland
  • Knatten, Anders Schau

Abstract

Techniques for checking memory operations are disclosed. A processing unit is accessed, comprising compute slices, control unit, local memory disambiguation units (LMDUs), and a global MDU (GMDU). Each slice includes an execution unit and is coupled to successor and predecessor slices. Each slice is coupled to an LMDU. Each LMDU is coupled to the GMDU. A first slice executes a first slice task. The task includes a load instruction and address. The slice issues the load to an LMDU, saving load information in a memory operation table (MOT). For a not fully serviced load instruction, the LMDU sends the load information to the GMDU, storing load information in a global MOT (GMOT). The GMOT detects address aliasing between the load address and a previously issued address saved in the GMOT. The GMOT forwards memory information from previously issued memory instructions to the MOT to satisfy the load instruction.

IPC Classes  ?

  • G06F 3/06 - Digital input from, or digital output to, record carriers

3.

LOCAL MEMORY DISAMBIGUATION FOR A PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number 19093385
Status Pending
Filing Date 2025-03-28
First Publication Date 2025-10-02
Owner Ascenium, Inc. (USA)
Inventor
  • Taylor, Jacob John Vorland
  • Knatten, Anders Schau

Abstract

A processing unit is accessed, comprising compute slices, a control unit, local memory disambiguation units (LMDUs), and memory system. Each slice includes an execution unit and is coupled to successor and predecessor slices. Each slice is coupled to an LMDU. The control unit distributes a first slice task to a first slice coupled to a first LMDU. The first slice executes the first task. The task includes a load instruction including a load address. The first slice issues the load instruction to the first LMDU. The issuing saves load information in a memory operation table (MOT) within the LMDU. The LMDU detects, based on the MOT, address aliasing between the load address and a store address of a previous store instruction. The MOT forwards store information from the previous store instruction. The store information satisfies one or more bytes of data required for the load instruction.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/46 - Multiprogramming arrangements

4.

COMPILER GENERATED HYPERBLOCKS IN A PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number 19053495
Status Pending
Filing Date 2025-02-14
First Publication Date 2025-08-21
Owner Ascenium, Inc. (USA)
Inventor
  • Aaser, Peter
  • Fredriksson, Hans Olle Viktor

Abstract

Techniques for parallel generation of blocks using a compiler are disclosed. A processing unit comprising compute slices, barrier register sets, a control unit, and a memory system is accessed. Each compute slice includes an execution unit and is coupled to other compute slices by a barrier register set. A compiler evaluates a compiled program that includes basic blocks, based on a control flow graph. A first hyperblock is created from at least two basic blocks. One or more branch instructions are replaced with skip instructions that direct instruction execution between basic blocks in the hyperblock. A first slice task is allocated to a first compute slice. A second slice task is allotted based on branch prediction. Pointers to the first compute slice and the second compute slice are initialized. The compiled program is executed, beginning with the first compute slice.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 8/41 - Compilation
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

5.

SEMANTIC ORDERING FOR PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number 18826260
Status Pending
Filing Date 2024-09-06
First Publication Date 2025-03-13
Owner Ascenium, Inc. (USA)
Inventor Taylor, Jacob John Vorland

Abstract

Techniques for managing compute slice tasks are disclosed. A processing unit comprising compute slices, load-store units (LSUs), a control unit, and a memory system is accessed. The compute slices are coupled. Each compute slice includes an LSU which is coupled to a predecessor LSU and a successor LSU. A compiled program is executed as the control unit distributes slice tasks to the compute slices for execution. A slice task, which includes a load instruction, is distributed to a current compute slice. The current compute slice can execute the slice task speculatively. A previously executed store instruction is committed to memory by a predecessor LSU. Address aliasing is checked between an address associated with the previously executed store instruction and the load address associated with the load instruction. The slice task running on the current compute slice can be cancelled when aliasing is detected.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

6.

SEMANTIC ORDERING FOR PARALLEL ARCHITECTURE WITH COMPUTE SLICES

      
Application Number US2024045453
Publication Number 2025/054375
Status In Force
Filing Date 2024-09-06
Publication Date 2025-03-13
Owner ASCENIUM, INC. (USA)
Inventor Taylor, Jacob John, Vorland

Abstract

Techniques for managing compute slice tasks are disclosed. A processing unit comprising compute slices, load-store units (LSUs), a control unit, and a memory system is accessed. The compute slices are coupled. Each compute slice includes an LSU which is coupled to a predecessor LSU and a successor LSU. A compiled program is executed as the control unit distributes slice tasks to the compute slices for execution. A slice task, which includes a load instruction, is distributed to a current compute slice. The current compute slice can execute the slice task speculatively. A previously executed store instruction is committed to memory by a predecessor LSU. Address aliasing is checked between an address associated with the previously executed store instruction and the load address associated with the load instruction. The slice task running on the current compute slice can be cancelled when aliasing is detected.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 12/0893 - Caches characterised by their organisation or structure

7.

PARALLEL PROCESSING ARCHITECTURE WITH BLOCK MOVE BACKPRESSURE

      
Application Number US2024044587
Publication Number 2025/049849
Status In Force
Filing Date 2024-08-30
Publication Date 2025-03-06
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for monitoring block moves in an array of compute elements and applying backpressure are disclosed. An array of compute elements is accessed. The array of compute elements is coupled to at least one data cache. The data cache provides memory' storage for the array of compute elements. Control for the array of compute elements is enabled by a stream of wide control words generated by the compiler. A load address and a store address comprising memory block move addresses are generated. The memory block move addresses point to memory' storage locations in the at least one data cache. Load buffers are coupled to the array of compute elements. The load buffers are located adjacent to at least one edge of the array of compute elements. A memory block move is executed using at least one of the load buffers, based on the memory block move addresses.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

8.

PARALLEL PROCESSING ARCHITECTURE WITH BLOCK MOVE SUPPORT

      
Application Number US2024039674
Publication Number 2025/024736
Status In Force
Filing Date 2024-07-26
Publication Date 2025-01-30
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing are disclosed. An array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is coupled to at least one data cache. The data cache provides memory storage for the array. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. A load address and a store address are generated. The load and the store addresses comprise memory block move addresses. The memory block move addresses point to memory storage locations in the data cache. A memory block move is executed, based on the memory block move addresses. The data for the memory block move is transferred outside of the array.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 15/173 - Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star or snowflake
  • G06F 12/084 - Multiuser, multiprocessor or multiprocessing cache systems with a shared cache
  • G06F 13/16 - Handling requests for interconnection or transfer for access to memory bus

9.

PARALLEL ARCHITECTURE WITH COMPILER-SCHEDULED COMPUTE SLICES

      
Application Number 18769478
Status Pending
Filing Date 2024-07-11
First Publication Date 2025-01-16
Owner Ascenium, Inc. (USA)
Inventor
  • Bastiansen, Tore Jahn
  • Aaser, Peter
  • Bø, Trond Hellem

Abstract

Techniques for task processing based on compiler-scheduled compute slices are disclosed. A processing unit comprising compute slices, barrier register sets, a control unit, and a memory system is accessed. Each compute slice includes an execution unit and is coupled to other compute slices by a barrier register set. A first slice task is distributed to a first compute slice. A second slice task is allotted to a second compute slice, based on a branch prediction logic. The second compute slice is coupled to the first by a first barrier register set. Pointers are initialized. A compiled program is executed, beginning at the first compute slice. The second slice task can be executed in parallel while a branch decision is being made. If the branch decision determines that the second slice task is not the next sequential slice task, results from the second compute slice are discarded.

IPC Classes  ?

  • G06F 9/52 - Program synchronisationMutual exclusion, e.g. by means of semaphores

10.

PARALLEL ARCHITECTURE WITH COMPILER-SCHEDULED COMPUTE SLICES

      
Application Number US2024037495
Publication Number 2025/015117
Status In Force
Filing Date 2024-07-11
Publication Date 2025-01-16
Owner ASCENIUM, INC. (USA)
Inventor
  • Bastiansen, Tore, Jahn
  • Aaser, Peter
  • Bø, Trond, Hellem

Abstract

Techniques for task processing based on compiler-scheduled compute slices are disclosed. A processing unit comprising compute slices, barrier register sets, a control unit, and a memory system is accessed. Each compute slice includes an execution unit and is coupled to other compute slices by a barrier register set. A first slice task is distributed to a first compute slice. A second slice task is allotted to a second compute slice, based on a branch prediction logic. The second compute slice is coupled to the first by a first barrier register set. Pointers are initialized. A compiled program is executed, beginning at the first compute slice. The second slice task can be executed in parallel while a branch decision is being made. If the branch decision determines that the second slice task is not the next sequential slice task, results from the second compute slice are discarded.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/52 - Program synchronisationMutual exclusion, e.g. by means of semaphores

11.

PARALLEL PROCESSING ARCHITECTURE WITH BLOCK MOVE BACKPRESSURE

      
Application Number 18820425
Status Pending
Filing Date 2024-08-30
First Publication Date 2024-12-19
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for monitoring block moves in an array of compute elements and applying backpressure are disclosed. An array of compute elements is accessed. The array of compute elements is coupled to at least one data cache. The data cache provides memory storage for the array of compute elements. Control for the array of compute elements is enabled by a stream of wide control words generated by the compiler. A load address and a store address comprising memory block move addresses are generated. The memory block move addresses point to memory storage locations in the at least one data cache. Load buffers are coupled to the array of compute elements. The load buffers are located adjacent to at least one edge of the array of compute elements. A memory block move is executed using at least one of the load buffers, based on the memory block move addresses.

IPC Classes  ?

  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]

12.

PARALLEL PROCESSING ARCHITECTURE WITH BLOCK MOVE SUPPORT

      
Application Number 18785026
Status Pending
Filing Date 2024-07-26
First Publication Date 2024-11-21
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing are disclosed. An array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is coupled to at least one data cache. The data cache provides memory storage for the array. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. A load address and a store address are generated. The load and the store addresses comprise memory block move addresses. The memory block move addresses point to memory storage locations in the data cache. A memory block move is executed, based on the memory block move addresses. The data for the memory block move is transferred outside of the array.

IPC Classes  ?

  • G06F 12/084 - Multiuser, multiprocessor or multiprocessing cache systems with a shared cache
  • G06F 9/54 - Interprogram communication

13.

PARALLEL PROCESSING HAZARD MITIGATION AVOIDANCE

      
Application Number US2024025310
Publication Number 2024/220740
Status In Force
Filing Date 2024-04-19
Publication Date 2024-10-24
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on hazard mitigation avoidance are disclosed. An array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operation hazard mitigation is enabled. The hazard mitigation is enabled by a control word tag. The control word tag supports memory access precedence information and is provided by the compiler at compile time. A hazardless memory access operation is executed. The hazardless memory access operation is determined by the compiler, and the hazardless memory access operation is designated by a unique set of precedence information contained in the tag. The tag is modified during runtime by hardware.

IPC Classes  ?

  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 8/41 - Compilation
  • G06F 12/0846 - Cache with multiple tag or data arrays being simultaneously accessible
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

14.

PARALLEL PROCESSING ARCHITECTURE FOR BRANCH PATH SUPPRESSION

      
Application Number US2024016978
Publication Number 2024/178275
Status In Force
Filing Date 2024-02-23
Publication Date 2024-08-29
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for a parallel processing architecture for branch path suppression are disclosed. An array of compute elements is accessed. Each element is known to a compiler and is coupled to its neighboring elements. Control for the elements is provided on a cycle- by -cycle basis. Control is enabled by a stream of wide control words generated by the compiler. The control includes a branch. A plurality of compute elements is mapped. The mapping distributes parallelized operations to the compute elements. The mapping is determined by the compiler. A column of compute elements is enabled to perform vertical data access suppression and a row of compute elements is enabled to perform horizontal data access suppression. Both sides of the branch are executed. The executing includes making a branch decision. Branch operation data accesses are suppressed, based on the branch decision and an invalid indication. The invalid indication is propagated among compute elements.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation

15.

PARALLEL PROCESSING HAZARD MITIGATION AVOIDANCE

      
Application Number 18640044
Status Pending
Filing Date 2024-04-19
First Publication Date 2024-08-08
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on hazard mitigation avoidance are disclosed. An array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operation hazard mitigation is enabled. The hazard mitigation is enabled by a control word tag. The control word tag supports memory access precedence information and is provided by the compiler at compile time. A hazardless memory access operation is executed. The hazardless memory access operation is determined by the compiler, and the hazardless memory access operation is designated by a unique set of precedence information contained in the tag. The tag is modified during runtime by hardware.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 12/0895 - Caches characterised by their organisation or structure of parts of caches, e.g. directory or tag array

16.

PARALLEL PROCESSING WITH HAZARD DETECTION AND STORE PROBES

      
Application Number US2024013611
Publication Number 2024/163521
Status In Force
Filing Date 2024-01-30
Publication Date 2024-08-08
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing using hazard detection and store probes are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Data to be stored by the array of compute elements is managed. The data to be stored is targeted to a data cache coupled to the array of compute elements. The managing includes detecting and mitigating memory hazards. Pending data cache accesses are probed for hazards. The examining comprises a store probe. Store data is committed to the data cache. The committing is based on a result of the store probe.

IPC Classes  ?

  • G06F 3/06 - Digital input from, or digital output to, record carriers
  • G06F 12/0802 - Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches

17.

PARALLEL PROCESSING ARCHITECTURE FOR BRANCH PATH SUPPRESSION

      
Application Number 18585156
Status Pending
Filing Date 2024-02-23
First Publication Date 2024-06-13
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for a parallel processing architecture for branch path suppression are disclosed. An array of compute elements is accessed. Each element is known to a compiler and is coupled to its neighboring elements. Control for the elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. The control includes a branch. A plurality of compute elements is mapped. The mapping distributes parallelized operations to the compute elements. The mapping is determined by the compiler. A column of compute elements is enabled to perform vertical data access suppression and a row of compute elements is enabled to perform horizontal data access suppression. Both sides of the branch are executed. The executing includes making a branch decision. Branch operation data accesses are suppressed, based on the branch decision and an invalid indication. The invalid indication is propagated among compute elements.

IPC Classes  ?

  • G06F 9/52 - Program synchronisationMutual exclusion, e.g. by means of semaphores

18.

PARALLEL PROCESSING WITH SWITCH BLOCK EXECUTION

      
Application Number US2023037172
Publication Number 2024/107390
Status In Force
Filing Date 2023-11-13
Publication Date 2024-05-23
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on parallel processing with switch block execution are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. A plurality of compute elements is initialized within the array with a switch statement. The switch statement is mapped into a primitive operation in each element of the plurality of compute elements. The initializing is based on a control word from the stream of control words. Each of the primitive operations is executed in an architectural cycle. A result is returned for the switch statement. The returning is determined by a decision variable.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation

19.

PARALLEL PROCESSING WITH HAZARD DETECTION AND STORE PROBES

      
Application Number 18426438
Status Pending
Filing Date 2024-01-30
First Publication Date 2024-05-23
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing using hazard detection and store probes are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Data to be stored by the array of compute elements is managed. The data to be stored is targeted to a data cache coupled to the array of compute elements. The managing includes detecting and mitigating memory hazards. Pending data cache accesses are probed for hazards. The examining comprises a store probe. Store data is committed to the data cache. The committing is based on a result of the store probe.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]

20.

PARALLEL PROCESSING USING HAZARD DETECTION AND MITIGATION

      
Application Number US2023036949
Publication Number 2024/107364
Status In Force
Filing Date 2023-11-07
Publication Date 2024-05-23
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing using hazard detection and mitigation are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operations are tagged with precedence information. The tagging is contained in the control words. The tagging is provided by the compiler at compile time. Memory access operations are monitored. The monitoring is based on the precedence information and a number of architectural cycles of the cycle-by- cycle basis. The tagging is augmented at run time, based on the monitoring. Memory access data is held before promotion, based on the monitoring.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 8/41 - Compilation

21.

PARALLEL PROCESSING WITH SWITCH BLOCK EXECUTION

      
Application Number 18388875
Status Pending
Filing Date 2023-11-13
First Publication Date 2024-03-07
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on parallel processing with switch block execution are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. A plurality of compute elements is initialized within the array with a switch statement. The switch statement is mapped into a primitive operation in each element of the plurality of compute elements. The initializing is based on a control word from the stream of control words. Each of the primitive operations is executed in an architectural cycle. A result is returned for the switch statement. The returning is determined by a decision variable.

IPC Classes  ?

  • G06F 12/0842 - Multiuser, multiprocessor or multiprocessing cache systems for multiprocessing or multitasking
  • G06F 12/0831 - Cache consistency protocols using a bus scheme, e.g. with bus monitoring or watching means

22.

PARALLEL PROCESSING ARCHITECTURE WITH MEMORY BLOCK TRANSFERS

      
Application Number US2023031460
Publication Number 2024/049859
Status In Force
Filing Date 2023-08-30
Publication Date 2024-03-07
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture with memory block transfers are disclosed. An array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. Control for the array is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. A control word from the stream of control words includes a source address, a target address, a block size, and a stride. Memory block transfer control logic is used. The memory block transfer logic is implemented outside of the array of compute elements. A memory block transfer is executed. The memory block transfer is initiated by a control word from the stream of wide control words. Data for the memory block transfer is moved independently from the array of compute elements.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

23.

Parallel processing using hazard detection and mitigation

      
Application Number 18387695
Grant Number 12493554
Status In Force
Filing Date 2023-11-07
First Publication Date 2024-02-29
Grant Date 2025-12-09
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing using hazard detection and mitigation are disclosed. An array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operations are tagged with precedence information. The tagging is contained in the control words. The tagging is provided by the compiler at compile time. Memory access operations are monitored. The monitoring is based on the precedence information and a number of architectural cycles of the cycle-by-cycle basis. The tagging is augmented at run time, based on the monitoring. Memory access data is held before promotion, based on the monitoring.

IPC Classes  ?

  • G06F 12/0846 - Cache with multiple tag or data arrays being simultaneously accessible
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

24.

PARALLEL PROCESSING ARCHITECTURE WITH BIN PACKING

      
Application Number US2023030769
Publication Number 2024/044150
Status In Force
Filing Date 2023-08-22
Publication Date 2024-02-29
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with bin packing are disclosed. An array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A plurality of compressed control words is generated by the compiler. The plurality of control words enables compute element operation and compute element memory access. The compressed control words are operationally sequenced. The compressed control words are linked by the compiler. Linking information is contained in at least one field of each of the compressed control words. The compressed control words are loaded into a control word cache coupled to the array of compute elements. The compressed control words are loaded into the control word cache in an operationally non-sequenced order. The plurality of compressed control words is ordered into an operationally sequenced execution order, based on the linking information.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 5/01 - Methods or arrangements for data conversion without changing the order or content of the data handled for shifting, e.g. justifying, scaling, normalising

25.

PARALLEL PROCESSING ARCHITECTURE WITH DUAL LOAD BUFFERS

      
Application Number US2023029057
Publication Number 2024/030351
Status In Force
Filing Date 2023-07-31
Publication Date 2024-02-08
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with dual load buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A first data cache is coupled to the array. The first data cache enables loading data to a first portion of the array. The first data cache supports an address space. A second data cache is coupled to the array. The second data cache enables loading data to a second portion of the array. The second data cache supports the address space. Instructions are executed within the array. Instructions executed within the first portion of the array of compute elements use data loaded from the first data cache, and instructions executed within the second portion of the array of compute elements use data loaded from the second data cache.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

26.

PARALLEL PROCESSING ARCHITECTURE WITH BIN PACKING

      
Application Number 18236442
Status Pending
Filing Date 2023-08-22
First Publication Date 2024-01-25
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with bin packing are disclosed. An array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A plurality of compressed control words is generated by the compiler. The plurality of control words enables compute element operation and compute element memory access. The compressed control words are operationally sequenced. The compressed control words are linked by the compiler. Linking information is contained in at least one field of each of the compressed control words. The compressed control words are loaded into a control word cache coupled to the array of compute elements. The compressed control words are loaded into the control word cache in an operationally non-sequenced order. The plurality of compressed control words is ordered into an operationally sequenced execution order, based on the linking information.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

27.

PARALLEL PROCESSING ARCHITECTURE WITH COUNTDOWN TAGGING

      
Application Number US2023027313
Publication Number 2024/015318
Status In Force
Filing Date 2023-07-11
Publication Date 2024-01-18
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with countdown tagging are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. A load operation is tagged with a countdown tag. Tagging is performed by the compiler, and the load operation is targeted to a memory system associated with the array of compute elements. The countdown tag comprises a time value. The time value is decremented as the load operation is being performed. The time value that is decremented is based on an architectural cycle. Countdown tag status is monitored by a control unit. The monitoring occurs as the load operation is performed. A load status is generated by the control unit, based on the monitoring. The load status allows compute element operation.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

28.

PARALLEL PROCESSING ARCHITECTURE WITH SPLIT CONTROL WORD CACHES

      
Application Number US2023026518
Publication Number 2024/006401
Status In Force
Filing Date 2023-06-29
Publication Date 2024-01-04
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for a parallel processing architecture with split control word caches are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A first control word cache is coupled to the array. The first control word cache enables loading control words to a first array portion. A second control word cache is coupled to the array. The second control word cache enables loading control words to a second array portion. The control words are split between the first and the second control word caches. The splitting is based on the constituency of the first and the second array portions. Instructions are executed within the array. Instructions executed within the first array portion use control words loaded from the first cache. Instructions executed within the second array portion use control words loaded from the second cache.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 12/0802 - Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches

29.

PARALLEL PROCESSING ARCHITECTURE WITH MEMORY BLOCK TRANSFERS

      
Application Number 18239770
Status Pending
Filing Date 2023-08-30
First Publication Date 2023-12-21
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture with memory block transfers are disclosed. An array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. Control for the array is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. A control word from the stream of control words includes a source address, a target address, a block size, and a stride. Memory block transfer control logic is used. The memory block transfer logic is implemented outside of the array of compute elements. A memory block transfer is executed. The memory block transfer is initiated by a control word from the stream of wide control words. Data for the memory block transfer is moved independently from the array of compute elements.

IPC Classes  ?

  • G06F 9/345 - Addressing or accessing the instruction operand or the result of multiple operands or results
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/54 - Interprogram communication

30.

PARALLEL PROCESSING ARCHITECTURE WITH DUAL LOAD BUFFERS

      
Application Number 18228001
Status Pending
Filing Date 2023-07-31
First Publication Date 2023-11-23
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with dual load buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A first data cache is coupled to the array. The first data cache enables loading data to a first portion of the array. The first data cache supports an address space. A second data cache is coupled to the array. The second data cache enables loading data to a second portion of the array. The second data cache supports the address space. Instructions are executed within the array. Instructions executed within the first portion of the array of compute elements use data loaded from the first data cache, and instructions executed within the second portion of the array of compute elements use data loaded from the second data cache.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 12/0842 - Multiuser, multiprocessor or multiprocessing cache systems for multiprocessing or multitasking

31.

PARALLEL PROCESSING OF MULTIPLE LOOPS WITH LOADS AND STORES

      
Application Number US2023021618
Publication Number 2023/220099
Status In Force
Filing Date 2023-05-10
Publication Date 2023-11-16
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing of multiple loops with loads and stores are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operations are tagged with precedence information. The tagging is contained in the control words and is implemented for loop operations. The tagging is provided by the compiler at compile time. Control word data is loaded for multiple, independent loops into the compute elements. The multiple, independent loops are executed. Memory is accessed based on the precedence information. The memory access includes loads and/or stores for data relating to the independent loops.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 8/41 - Compilation

32.

PARALLEL PROCESSING ARCHITECTURE WITH COUNTDOWN TAGGING

      
Application Number 18220331
Status Pending
Filing Date 2023-07-11
First Publication Date 2023-11-02
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing based on a parallel processing architecture with countdown tagging are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. A load operation is tagged with a countdown tag. Tagging is performed by the compiler, and the load operation is targeted to a memory system associated with the array of compute elements. The countdown tag comprises a time value. The time value is decremented as the load operation is being performed. The time value that is decremented is based on an architectural cycle. Countdown tag status is monitored by a control unit. The monitoring occurs as the load operation is performed. A load status is generated by the control unit, based on the monitoring. The load status allows compute element operation.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 11/30 - Monitoring
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

33.

PARALLEL PROCESSING ARCHITECTURE WITH SPLIT CONTROL WORD CACHES

      
Application Number 18215866
Status Pending
Filing Date 2023-06-29
First Publication Date 2023-10-26
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for a parallel processing architecture with split control word caches are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. A first control word cache is coupled to the array. The first control word cache enables loading control words to a first array portion. A second control word cache is coupled to the array. The second control word cache enables loading control words to a second array portion. The control words are split between the first and the second control word caches. The splitting is based on the constituency of the first and the second array portions. Instructions are executed within the array. Instructions executed within the first array portion use control words loaded from the first cache. Instructions executed within the second array portion use control words loaded from the second cache.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

34.

AUTONOMOUS COMPUTE ELEMENT OPERATION USING BUFFERS

      
Application Number US2023015735
Publication Number 2023/183279
Status In Force
Filing Date 2023-03-21
Publication Date 2023-09-28
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on autonomous compute element operation using buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control is provided for the array of compute elements on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. An autonomous operation buffer is loaded with at least two operations contained in control words. The autonomous operation buffer is integrated in a compute element. A compute element operation counter coupled to the autonomous operation buffer is set. The compute element operation counter is integrated in the compute element. The at least two operations are executed using the autonomous operation buffer and the compute element operation counter. The operations complete autonomously from direct compiler control.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 8/41 - Compilation
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 7/57 - Arithmetic logic units [ALU], i.e. arrangements or devices for performing two or more of the operations covered by groups or for performing logical operations

35.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH OUT-OF-ORDER RESOLUTION

      
Application Number US2023014863
Publication Number 2023/172660
Status In Force
Filing Date 2023-03-09
Publication Date 2023-09-14
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a highly parallel processing architecture with out-of-order resolution are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. The array of compute elements is coupled to supporting logic and to memory, which, along with the array of compute elements, comprise compute hardware. A set of directions is provided to the hardware, through a control word generated by the compiler, for compute element operation. The set of directions is augmented with data access ordering information. The data access ordering is performed by the hardware. A compiled task is executed on the array of compute elements, based on the set of directions that was augmented.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation

36.

PARALLEL PROCESSING OF MULTIPLE LOOPS WITH LOADS AND STORES

      
Application Number 18195407
Status Pending
Filing Date 2023-05-10
First Publication Date 2023-09-07
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for parallel processing of multiple loops with loads and stores are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array. Control for the compute elements is provided on a cycle-by-cycle basis. Control is enabled by a stream of wide control words generated by the compiler. Memory access operations are tagged with precedence information. The tagging is contained in the control words and is implemented for loop operations. The tagging is provided by the compiler at compile time. Control word data is loaded for multiple, independent loops into the compute elements. The multiple, independent loops are executed. Memory is accessed based on the precedence information. The memory access includes loads and/or stores for data relating to the independent loops.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

37.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH OUT-OF-ORDER RESOLUTION

      
Application Number 18119315
Status Pending
Filing Date 2023-03-09
First Publication Date 2023-08-31
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a highly parallel processing architecture with out-of-order resolution are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. The array of compute elements is coupled to supporting logic and to memory, which, along with the array of compute elements, comprise compute hardware. A set of directions is provided to the hardware, through a control word generated by the compiler, for compute element operation. The set of directions is augmented with data access ordering information. The data access ordering is performed by the hardware. A compiled task is executed on the array of compute elements, based on the set of directions that was augmented.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 8/41 - Compilation

38.

AUTONOMOUS COMPUTE ELEMENT OPERATION USING BUFFERS

      
Application Number 18124115
Status Pending
Filing Date 2023-03-21
First Publication Date 2023-07-13
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on autonomous compute element operation using buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control is provided for the array of compute elements on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. An autonomous operation buffer is loaded with at least two operations contained in control words. The autonomous operation buffer is integrated in a compute element. A compute element operation counter coupled to the autonomous operation buffer is set. The compute element operation counter is integrated in the compute element. The at least two operations are executed using the autonomous operation buffer and the compute element operation counter. The operations complete autonomously from direct compiler control.

IPC Classes  ?

39.

COMPUTE ELEMENT PROCESSING USING CONTROL WORD TEMPLATES

      
Application Number US2022053912
Publication Number 2023/129491
Status In Force
Filing Date 2022-12-23
Publication Date 2023-07-06
Owner ASCENIUM, INC. (USA)
Inventor
  • Hristodorescu, Ionut
  • Foley, Peter

Abstract

Techniques for task processing based on compute element processing using control word templates are disclosed. One or more control word templates are generated for use in a two-dimensional array of compute elements. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Each control word template designates a topological set of compute elements from the array of compute elements. The one or more control word templates are customized with a specific set of compute element operations. The one or more control word templates that were customized are stored. The specific set of compute element operations is executed on the topological set of compute elements. The one or more control word templates that were stored are reused. The one or more control word templates that were stored are modified and executed using compute elements.

IPC Classes  ?

  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 8/41 - Compilation
  • G06F 15/173 - Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star or snowflake
  • G06F 7/57 - Arithmetic logic units [ALU], i.e. arrangements or devices for performing two or more of the operations covered by groups or for performing logical operations

40.

COMPUTE ELEMENT PROCESSING USING CONTROL WORD TEMPLATES

      
Application Number 18087865
Status Pending
Filing Date 2022-12-23
First Publication Date 2023-04-27
Owner ASCENIUM, INC. (USA)
Inventor
  • Hristodorescu, Ionut
  • Foley, Peter

Abstract

Techniques for task processing based on compute element processing using control word templates are disclosed. One or more control word templates are generated for use in a two-dimensional array of compute elements. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Each control word template designates a topological set of compute elements from the array of compute elements. The one or more control word templates are customized with a specific set of compute element operations. The one or more control word templates that were customized are stored. The specific set of compute element operations is executed on the topological set of compute elements. The one or more control word templates that were stored are reused. The one or more control word templates that were stored are modified and executed using compute elements.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 40/186 - Templates

41.

LOAD LATENCY AMELIORATION USING BUNCH BUFFERS

      
Application Number US2022046210
Publication Number 2023/064230
Status In Force
Filing Date 2022-10-11
Publication Date 2023-04-20
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on load latency amelioration using bunch buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. Sets of control word bits are loaded into buffers. Each buffer is associated with and coupled to a unique compute element within the array of compute elements. The sets of control word bits provide operational control for the compute element with which it is associated. Operations are executed within the array of elements. The operations are based on a selected set of control word bits which comprise a control word bunch.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 15/78 - Architectures of general purpose stored program computers comprising a single central processing unit
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt

42.

PARALLEL PROCESSING ARCHITECTURE USING DISTRIBUTED REGISTER FILES

      
Application Number US2022033510
Publication Number 2023/018477
Status In Force
Filing Date 2022-06-15
Publication Date 2023-02-16
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture using distributed register files are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is controlled on a cycle-by-cycle basis. The controlling is enabled by a stream of wide control words generated by the compiler. Virtual registers are mapped to a plurality of physical register files distributed among one or more of the compute elements. Virtual registers are represented by the compiler. The mapping is performed by the compiler. A broadcast write operation is enabled to two or more of the physical register files. Operations contained in the control words are executed. Operations are enabled by at least one of the distributed physical register files. Implementation in separate compute elements enables parallel operation processing.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

43.

PARALLEL PROCESSING ARCHITECTURE FOR ATOMIC OPERATIONS

      
Application Number US2022039234
Publication Number 2023/014759
Status In Force
Filing Date 2022-08-03
Publication Date 2023-02-09
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing in a parallel processing architecture for atomic operations are disclosed. A two-dimensional array of compute elements is accessed, where each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. At least one of the control words involves an operation requiring at least one additional operation. A bit of the control word is set, where the bit indicates a multicycle operation. The control word is executed, on at least one compute element within the array of compute elements, based on the bit. The multicycle operation comprises a read-modify -write operation.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 8/41 - Compilation
  • G06F 13/40 - Bus structure
  • G06F 13/16 - Handling requests for interconnection or transfer for access to memory bus

44.

LOAD LATENCY AMELIORATION USING BUNCH BUFFERS

      
Application Number 17963226
Status Pending
Filing Date 2022-10-11
First Publication Date 2023-02-02
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on load latency amelioration using bunch buffers are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. Sets of control word bits are loaded into buffers. Each buffer is associated with and coupled to a unique compute element within the array of compute elements. The sets of control word bits provide operational control for the compute element with which it is associated. Operations are executed within the array of elements. The operations are based on a selected set of control word bits which comprise a control word bunch.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/54 - Interprogram communication

45.

PARALLEL PROCESSING ARCHITECTURE WITH DISTRIBUTED REGISTER FILES

      
Application Number US2022030800
Publication Number 2022/251272
Status In Force
Filing Date 2022-05-25
Publication Date 2022-12-01
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture with distributed register files are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is controlled on a cycle-by-cycle basis. The controlling is enabled by a stream of wide, variable length, control words generated by the compiler. Virtual registers are mapped to a plurality of physical register files distributed among one or more of the compute elements. Virtual registers are represented by the compiler. The mapping is performed by the compiler. A broadcast write operation is enabled to two or more of the physical register files. Operations contained in the control words are executed. Operations are enabled by at least one of the distributed physical register files. Implementation in separate compute elements enables parallel operation processing.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 8/41 - Compilation

46.

PARALLEL PROCESSING ARCHITECTURE FOR ATOMIC OPERATIONS

      
Application Number 17879827
Status Pending
Filing Date 2022-08-03
First Publication Date 2022-11-24
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing in a parallel processing architecture for atomic operations are disclosed. A two-dimensional array of compute elements is accessed, where each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide control words generated by the compiler. At least one of the control words involves an operation requiring at least one additional operation. A bit of the control word is set, where the bit indicates a multicycle operation. The control word is executed, on at least one compute element within the array of compute elements, based on the bit. The multicycle operation comprises a read-modify-write operation.

IPC Classes  ?

  • G06F 9/52 - Program synchronisationMutual exclusion, e.g. by means of semaphores
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt

47.

PARALLEL PROCESSING ARCHITECTURE USING DISTRIBUTED REGISTER FILES

      
Application Number 17840689
Status Pending
Filing Date 2022-06-15
First Publication Date 2022-09-29
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture using distributed register files are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is controlled on a cycle-by-cycle basis. The controlling is enabled by a stream of wide control words generated by the compiler. Virtual registers are mapped to a plurality of physical register files distributed among one or more of the compute elements. Virtual registers are represented by the compiler. The mapping is performed by the compiler. A broadcast write operation is enabled to two or more of the physical register files. Operations contained in the control words are executed. Operations are enabled by at least one of the distributed physical register files. Implementation in separate compute elements enables parallel operation processing.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt

48.

PARALLEL PROCESSING ARCHITECTURE USING SPECULATIVE ENCODING

      
Application Number US2022021838
Publication Number 2022/204450
Status In Force
Filing Date 2022-03-25
Publication Date 2022-09-29
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for program execution in a parallel processing architecture using speculative encoding are disclosed. A two-dimensional array of compute elements is accessed, where each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, control words generated by the compiler. Two or more operations are coalesced into a control word, where the control word includes a branch decision and operations associated with the branch decision. The coalesced control word includes speculatively encoded operations for at least two possible branch paths. The at least two possible branch paths generate independent side effects. Operations associated with the branch decision that are not indicated by the branch decision are suppressed.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 7/575 - Basic arithmetic logic units, i.e. devices selectable to perform either addition, subtraction or one of several logical operations, using, at least partially, the same circuitry
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 8/41 - Compilation

49.

Parallel processing architecture with distributed register files

      
Application Number 17752898
Grant Number 12578991
Status In Force
Filing Date 2022-05-25
First Publication Date 2022-09-15
Grant Date 2026-03-17
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing based on a parallel processing architecture with distributed register files are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. The array of compute elements is controlled on a cycle-by-cycle basis. The controlling is enabled by a stream of wide, variable length, control words generated by the compiler. Virtual registers are mapped to a plurality of physical register files distributed among one or more of the compute elements. Virtual registers are represented by the compiler. The mapping is performed by the compiler. A broadcast write operation is enabled to two or more of the physical register files. Operations contained in the control words are executed. Operations are enabled by at least one of the distributed physical register files. Implementation in separate compute elements enables parallel operation processing.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

50.

PARALLEL PROCESSING ARCHITECTURE USING SPECULATIVE ENCODING

      
Application Number 17704056
Status Pending
Filing Date 2022-03-25
First Publication Date 2022-07-07
Owner Ascenium, Inc. (USA)
Inventor Foley, Peter

Abstract

Techniques for program execution in a parallel processing architecture using speculative encoding are disclosed. A two-dimensional array of compute elements is accessed, where each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, control words generated by the compiler. Two or more operations are coalesced into a control word, where the control word includes a branch decision and operations associated with the branch decision. The coalesced control word includes speculatively encoded operations for at least two possible branch paths. The at least two possible branch paths generate independent side effects. Operations associated with the branch decision that are not indicated by the branch decision are suppressed.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

51.

HIGHLY PARALLEL PROCESSING ARCHITECTURE USING DUAL BRANCH EXECUTION

      
Application Number US2021063441
Publication Number 2022/132858
Status In Force
Filing Date 2021-12-15
Publication Date 2022-06-23
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing in a highly parallel processing architecture using dual branch execution are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, control words generated by the compiler. The control includes a branch. Two sides of the branch in the array are executed while waiting for a branch decision to be acted upon by control logic. The branch decision is based on computation results in the array. Data produced by a taken branch path is promoted. Results from a side of the branch not indicated by the branch decision are ignored or invalidated.

IPC Classes  ?

  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 8/41 - Compilation
  • G06F 15/78 - Architectures of general purpose stored program computers comprising a single central processing unit

52.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH COMPILER

      
Application Number US2021059304
Publication Number 2022/104176
Status In Force
Filing Date 2021-11-15
Publication Date 2022-05-19
Owner ASCENIUM, INC. (USA)
Inventor
  • Harboe, Øyvind
  • Bastiansen, Tore
  • Foley, Peter

Abstract

Techniques for task processing using a highly parallel processing architecture with a compiler are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. A set of directions is provided to the hardware, through a control word generated by the compiler, for compute element operation and memory access precedence. The set of directions enables the hardware to properly sequence compute element results. The set of directions controls data movement for the array of compute elements. A compiled task is executed on the array of compute elements, based on the set of directions. The compute element results are generated in parallel in the array, and the compute element results are ordered independently from control word arrival at each compute element.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode

53.

PARALLEL PROCESSING ARCHITECTURE WITH BACKGROUND LOADS

      
Application Number US2021054889
Publication Number 2022/081784
Status In Force
Filing Date 2021-10-14
Publication Date 2022-04-21
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing using a parallel processing architecture with background loads are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. Operation of the array is paused. The pausing occurs while a memory system continues operation. A bus coupling the array is repurposed. The repurposing couples one or more compute elements in the array to the memory system. A memory system operation is enabled during the pausing. Data is transferred from the memory system to the array of compute elements using the bus that was repurposed. The data from the memory system is transferred to scratchpad memory in the one or more compute elements within the two- dimensional array. The scratchpad memory provides operand storage. The data is tagged. The tagging guides the transferring to a particular compute element.

IPC Classes  ?

  • G06Q 10/06 - Resources, workflows, human or project managementEnterprise or organisation planningEnterprise or organisation modelling
  • G06Q 10/10 - Office automationTime management

54.

HIGHLY PARALLEL PROCESSING ARCHITECTURE USING DUAL BRANCH EXECUTION

      
Application Number 17551276
Status Pending
Filing Date 2021-12-15
First Publication Date 2022-04-07
Owner ASCENIUM, INC, (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing in a highly parallel processing architecture using dual branch execution are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, control words generated by the compiler. The control includes a branch. Two sides of the branch in the array are executed while waiting for a branch decision to be acted upon by control logic. The branch decision is based on computation results in the array. Data produced by a taken branch path is promoted. Results from a side of the branch not indicated by the branch decision are ignored or invalidated.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 7/575 - Basic arithmetic logic units, i.e. devices selectable to perform either addition, subtraction or one of several logical operations, using, at least partially, the same circuitry
  • G06N 3/04 - Architecture, e.g. interconnection topology

55.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH SHALLOW PIPELINE

      
Application Number US2021048964
Publication Number 2022/055792
Status In Force
Filing Date 2021-09-03
Publication Date 2022-03-17
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing using a highly parallel processing architecture with a shallow pipeline are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, microcode control words generated by the compiler. Relevant portions of the control word are stored within a cache associated with the array of compute elements. The control words are decompressed. The decompressing occurs cycle-by-cycle out of the cache over multiple cycles. A compiled task is executed on the array of compute elements, based on the decompressing. Simultaneous execution of two or more potential compiled task outcomes is provided.

IPC Classes  ?

  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/28 - Enhancement of operational speed, e.g. by using several microcontrol devices operating in parallel
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead

56.

PARALLEL PROCESSING ARCHITECTURE WITH BACKGROUND LOADS

      
Application Number 17500990
Status Pending
Filing Date 2021-10-14
First Publication Date 2022-03-10
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing using a parallel processing architecture with background loads are disclosed. A two-dimensional array of compute elements is accessed. Each compute element is known to a compiler and is coupled to its neighboring compute elements. Operation of the array is paused. The pausing occurs while a memory system continues operation. A bus coupling the array is repurposed. The repurposing couples one or more compute elements in the array to the memory system. A memory system operation is enabled during the pausing. Data is transferred from the memory system to the array of compute elements using the bus that was repurposed. The data from the memory system is transferred to scratchpad memory in the one or more compute elements within the two-dimensional array. The scratchpad memory provides operand storage. The data is tagged. The tagging guides the transferring to a particular compute element.

IPC Classes  ?

  • G06F 13/16 - Handling requests for interconnection or transfer for access to memory bus
  • G06F 13/40 - Bus structure

57.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH SHALLOW PIPELINE

      
Application Number 17465949
Status Pending
Filing Date 2021-09-03
First Publication Date 2022-03-10
Owner ASCENIUM, INC. (USA)
Inventor Foley, Peter

Abstract

Techniques for task processing using a highly parallel processing architecture with a shallow pipeline are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. Control for the array of compute elements is provided on a cycle-by-cycle basis. The control is enabled by a stream of wide, variable length, microcode control words generated by the compiler. Relevant portions of the control word are stored within a cache associated with the array of compute elements. The control words are decompressed. The decompressing occurs cycle-by-cycle out of the cache over multiple cycles. A compiled task is executed on the array of compute elements, based on the decompressing. Simultaneous execution of two or more potential compiled task outcomes is provided.

IPC Classes  ?

  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors

58.

HIGHLY PARALLEL PROCESSING ARCHITECTURE WITH COMPILER

      
Application Number 17526003
Status Pending
Filing Date 2021-11-15
First Publication Date 2022-03-10
Owner Ascenium, Inc. (USA)
Inventor
  • Harboe, Øyvind
  • Bastiansen, Tore
  • Foley, Peter

Abstract

Techniques for task processing using a highly parallel processing architecture with a compiler are disclosed. A two-dimensional array of compute elements is accessed. Each compute element within the array of compute elements is known to a compiler and is coupled to its neighboring compute elements within the array of compute elements. A set of directions is provided to the hardware, through a control word generated by the compiler, for compute element operation and memory access precedence. The set of directions enables the hardware to properly sequence compute element results. The set of directions controls data movement for the array of compute elements. A compiled task is executed on the array of compute elements, based on the set of directions. The compute element results are generated in parallel in the array, and the compute element results are ordered independently from control word arrival at each compute element.

IPC Classes  ?

  • G06F 9/48 - Program initiatingProgram switching, e.g. by interrupt
  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]

59.

Method of placement and routing in a reconfiguration of a dynamically reconfigurable processor

      
Application Number 13832005
Grant Number 09633160
Status In Force
Filing Date 2013-03-15
First Publication Date 2014-05-01
Grant Date 2017-04-25
Owner
  • ASCENIUM HOLDING AS (Norway)
  • ASCENIUM INC (USA)
Inventor Mykland, Robert Keith

Abstract

A method and system are provided for deriving a resultant compiled software code with increased compatibility for placement and routing of a dynamically reconfigurable processor.

IPC Classes  ?

60.

Method and system adapted for converting software constructs into resources for implementation by a dynamically reconfigurable processor

      
Application Number 13722831
Grant Number 09304770
Status In Force
Filing Date 2012-12-20
First Publication Date 2014-01-09
Grant Date 2016-04-05
Owner
  • ASCENIUM HOLDING AS (Norway)
  • ASCENIUM INC (USA)
Inventor Mykland, Robert Keith

Abstract

A method and system are provided for deriving a resultant software code from an originating ordered list of instructions that does not include overlapping branch logic. The method may include deriving a plurality of unordered software constructs from a sequence of processor instructions; associating software constructs in accordance with an original logic of the sequence of processor instructions; determining and resolving memory precedence conflicts within the associated plurality of software constructs; resolving forward branch logic structures into conditional logic constructs; resolving back branch logic structures into loop logic constructs; and/or applying the plurality of unordered software constructs in a programming operation by a parallel execution logic circuitry. The resultant plurality of unordered software constructs may be converted into programming reconfigurable logic, computers or processors, and also by means of a computer network or an electronics communications network.

IPC Classes  ?

  • G06F 7/00 - Methods or arrangements for processing data by operating upon the order or content of the data handled
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/44 - Arrangements for executing specific programs

61.

System and method for performing a branch object conversion to program configurable logic circuitry

      
Application Number 13493962
Grant Number 09158544
Status In Force
Filing Date 2012-06-11
First Publication Date 2013-06-06
Grant Date 2015-10-13
Owner
  • ASCENIUM HOLDING AS (Norway)
  • ASCENIUM INC (USA)
Inventor Mykland, Robert Keith

Abstract

A method and system are provided for deriving a resultant software code from an originating ordered list of instructions that does not include overlapping branch logic. The method may include deriving a plurality of unordered software constructs from a sequence of processor instructions; associating software constructs in accordance with an original logic of the sequence of processor instructions; determining and resolving memory precedence conflicts within the associated plurality of software constructs; resolving forward branch logic structures into conditional logic constructs; resolving back branch logic structures into loop logic constructs; and/or applying the plurality of unordered software constructs in a programming operation by a parallel execution logic circuitry. The resultant plurality of unordered software constructs may be converted into programming reconfigurable logic, computers or processors, and also by means of a computer network or an electronics communications network.

IPC Classes  ?

  • G06F 9/50 - Allocation of resources, e.g. of the central processing unit [CPU]
  • G06F 9/30 - Arrangements for executing machine instructions, e.g. instruction decode
  • G06F 9/38 - Concurrent instruction execution, e.g. pipeline or look ahead
  • G06F 9/45 - Compilation or interpretation of high level programme languages

62.

Configurable circuit array

      
Application Number 13301763
Grant Number 10089277
Status In Force
Filing Date 2011-11-21
First Publication Date 2012-12-27
Grant Date 2018-10-02
Owner
  • ASCENIUM HOLDING AS (Norway)
  • ASCENIUM INC (USA)
Inventor Mykland, Robert Keith

Abstract

A method and system are provided for configurable computation and data processing. A logical processor includes an array of logic elements. The processor may be a combinatorial circuit that can be applied to modify computational aspects of an array of reconfigurable circuits. A memory stores a plurality of instructions, each instruction including an instruction-fetch data portion and an output data transfer data portion. One or more memory controllers are coupled to the memory and receive instructions and/or output data from the memory. A back buffer is coupled with the memory controller and receives instructions from the memory controller. The back buffer sequentially asserts each received instruction upon one or more memory controllers. The memory controllers transfer data received from the memory to a target, such as an array of reconfigurable logic circuits that are optionally coupled to the memory, the back buffer, and one or more additional memory controllers.

IPC Classes  ?

  • G06F 15/78 - Architectures of general purpose stored program computers comprising a single central processing unit