Geneve video wait states

From Ninerpedia
Revision as of 15:03, 20 October 2011 by Mizapf (talk | contribs)
Jump to navigation Jump to search

The Geneve video wait state handling is controlled by the Gate array chip on the board. Unfortunately, there are no public specifications on that custom chip, so all we can do is to try by experiment.

Video wait states are created after the access has occured. On that occasion, 14 wait states are created, presumably by a counter in the Gate array, pulling down the READY line for that period of time. Before this, one or two wait states are inserted since the video access is an external one.

The long wait state period starts after the access, which causes this very access to terminate, and then to start the period for the next command. This means that those wait states are completely ineffective in the following case:

 // Assume workspace at F000
 F040    MOVB @VDPRD,R3
 F044    NOP
 F046    DEC  R1
 F048    JNE  F040

The instruction at address F040 requires 5 cycles: get MOVB @,R3, get the source address, read from the source address, 1 WS for this external access, write to register 3. Then the next command is read (NOP) which is in the on-chip RAM (check the addresses). For that reason, the READY pin is ignored, which has been pulled low in the meantime. NOP takes three cycles, as does DEC R1, still ignoring the low READY. JNE also takes three cycles. If R1 is not 0, we jump back, and the command is executed again.

Obviously, the counter is now reset, because we measure that the command again takes only 5 cycles, despite the fact that the 14 WS from last iteration are not yet over.

However, if we add an access to SRAM, we will notice a delay:

 // Assume workspace at F000
 F040    MOVB @VDPRD,R3     *  5 cycles
 F044    MOVB @SRAM,R4      * 15 cycles (4 cycles + 11 WS)
 F048    DEC  R1            *  3 cycles
 F04A    JNE  F040          *  3 cycles

But we only get 11 WS, not 14. This is reasonable if we remember that the access to SRAM does not occur immediately after the VDP access but after writing to R3 and getting the next command. This can be shown more easily in the following way. ADDR is the constant, symbolizing the memory location; *ADDR means an access to the memory location at address ADDR, < is reading, > is writing. PC is the program counter. The opcode, which is read in the first cycle, contains information about the types of source and destination locations, as well as the register numbers.

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
MOVB @,R3 VDPRD wait < *VDPRD > R3
MOVB @,R4 SRAM wait wait wait wait wait wait wait wait wait wait wait < *SRAM > R3
DEC R1 < R1 > R1
JNE < PC > PC

There are several activities going on while the wait states are active "in the background", as symbolized by the blue background. Only when the next access to external memory occurs the wait states become effective. As said above, the read operations only terminate when the READY line gets high again, so we put the termination of the read access in the last box. You should keep in mind that the read access starts right after the last activity.

We can spend so much time inside the on-chip address space that all wait states pass without effect. The following code is executed at the same speed, whether the video wait state bit is set of not:

 // Assume workspace at F000
 F040    MOVB @VDPRD,R3 
         NOP
         DEC  @>F006
         DEC  @>F006
         MOVB @SRAM,R4
         DEC  R1
         JNE  F040

as you can see here:

1 2 3 4 5
MOVB @,R3 VDPRD wait < *VDPRD > R3
NOP < PC > PC
DEC @ F006 < *F006 > *F006
DEC @ F006 < *F006 > *F006
MOVB @,R4 SRAM < *SRAM > R3
DEC R1 < R1 > R1
JNE < PC > PC

Now an example on writing to VDP.

F040   MOVB R3,@>VDPWD
       NOP
       NOP
       NOP
       MOVB @SRAM,@>F000
       DEC  R1
       JNE  F040

Again as a diagram. Note that box 5 in the first line is not backgrounded: The READY line must be high, or the command will not be completed. We assume that the data are written to VDPWD in box 4 already.

1 2 3 4 5 6 7 8 9
MOVB R3,@ < R3 VDPWD > *VDPWD wait
NOP < PC > PC
NOP < PC > PC
NOP < PC > PC
MOVB @,@ SRAM wait wait wait wait < *SRAM F000 > *F000
DEC R1 < R1 > R1
JNE < PC > PC

There is one peculiarity: We now have 15 wait states, not 14. Indeed, as my measurements prove, the complete code requires 29 cycles, and by using commands in the on-chip RAM as shown in the previous example one can see that the second MOVB line requires 9 cycles, including 4 wait states. As it seems, video writing requires one wait state more than reading. This may be a design issue of the Gate array, but it won't cause any trouble.

Finally we should look at the case when we get wait states by different reasons. Assume we set up the system to use video wait states and extra wait states.

F040   MOVB R3,@>VDPWD
       NOP
       NOP
       DEC  @>F000
       MOVB @SRAM,R4
       DEC  R1
       JNE  F040
1 2 3 4 5 6 7 8 9
MOVB R3,@ < R3 VDPWD > *VDPWD wait wait
NOP < PC > PC
NOP < PC > PC
DEC @ F000 < *F000 > *F000
MOVB @,R4 SRAM wait wait wait < *SRAM > R4
DEC R1 < R1 > R1
JNE < PC > PC

First thing to notice is that the first line gets another wait state, since we have requested additional wait states, and VDPWD is an external address. Also, we have 15 wait states from video (blue).

What happens to the additional wait states for the SRAM access in line 5? They occur at the same time as the video wait states, and therefore are not effective. As it seems, the counters for the wait states do not add on each other. The SRAM access is still delayed by 3 wait states (not 2) because there is one wait state left from the video access. After this one is over, the access occurs.

There are some unsolved questions, unfortunately. They are not really a problem, but they make it difficult to understand the actual implementation and the way how this could be used within an emulation.

  • Why do VDP write accesses cause 1 more wait state than read accesses?
  • If we put the code in SRAM and access video, the following command should be delayed by 14 WS. However, it turns out to require 15 WS.
F040   MOVB @>VDPRD,R3
       NOP
       MOVB @SRAM,R4
       DEC  R1
       JNE  F040

If this code is executed in the on-chip RAM it takes 27 cycles. When executed in SRAM it requires 41 cycles. This seems at first a good number, as we said reading causes 14 wait states. However, this ignores the fact that one wait state occurs while writing to R3, so we should only have 13 effective WS.

So we must assume 15 WS. We know that cause comes before effect: the Gate array cannot "know" after line 1 that it must set the counter 1 higher. On the other hand, the Gate array cannot tell whether the memory access is a data or an instruction acquisition. It could, using the IAQ processor line, but this one (you guess) is not connected to the Gate array.

Maybe there is an effect from the processor's instruction prefetch capability. As I already said, the TMS9995 is highly efficient in its usage of clock cycles, which is only possible when it can get the next instruction during internal operations. The additional wait state is probably be caused by the fact that external accesses are still blocked by the READY line, so one more cycle may be required.