Associative remote viewing has drawn attention after extensive testing by Greg Kolodziejzyk, who ran 5,677 trials over 13 years to assess its value. Researchers and curious minds study this practice to see if a percipient can describe a target image tied to a specific future occurrence.
The method asks a participant to focus on a sealed image that will be revealed only when an actual future event happens. Practitioners aim to bridge present awareness with what will come using anomalous cognition.
Many studies aim to find if this approach yields reliable, actionable data before an event takes place. The field mixes careful trials, statistical checks, and an open question about how the mind may access information beyond linear time.
Key Takeaways
- Greg Kolodziejzyk completed 5,677 trials over 13 years testing ARV methods.
- The practice asks a percipient to describe a sealed image linked to an upcoming event.
- Researchers seek evidence that the mind can access information about a future occurrence.
- ARV blends controlled testing with concepts of anomalous cognition.
- The technique aims to produce reliable signals before a target event is revealed.
Understanding the Basics of Associative Remote Viewing
In 1984, Russell Targ and Harold Puthoff formalized a protocol that linked visual targets to binary outcomes. This design aimed to tighten control in studies of precognition and make results easier to measure.
Defining the Concept
How it works: Two distinct images are paired with two possible outcomes. A percipient describes one image without knowing which outcome will occur.
- Protocol origin: Created to improve reliability in experimental work.
- Binary mapping: Visual targets act as proxies for yes/no or A/B choices.
- Bias reduction: Separation between percipient and outcome lowers conscious influence.

The Goal of Precognition
The main aim is clear: obtain repeatable signals that can guide decisions where standard forecasts struggle. Practitioners use this structured approach to keep impressions objective and testable.
“This method turns vague impressions into measurable matches, improving the chance of consistent results.”
Short sessions, strict controls, and scoring systems help researchers assess hits versus misses. Over time, these elements seek to raise the success rate of arv and related protocols.
Historical Foundations of Remote Viewing Experiments
A set of high-profile experiments in 1982 put unconventional perception techniques under financial scrutiny.
In 1982, Keith Harary and Russell Targ ran a notable remote viewing experiment that forecast silver futures for nine straight weeks. The trial produced more than $100,000 in profit and drew intense attention to the method.
That same year Harold E. Puthoff led a parallel experiment that returned about $25,000 in 30 days. These early results helped legitimize small-scale trials and inspired systematic studies in academic outlets like the Journal of Parapsychology.
The outcomes offered raw data that justified more rigorous research and replication. Scholars used these reports to refine protocols and improve scoring in later studies.

| Year | Lead Researcher | Study Type | Reported Results |
|---|---|---|---|
| 1982 | Keith Harary & Russell Targ | remote viewing experiment | $100,000+ profit (9 weeks) |
| 1982 | Harold E. Puthoff | experiment | $25,000 profit (30 days) |
| 1980s | Various teams | studies published | Peer reports in journal parapsychology |
For readers curious about related abilities, see a concise overview of clairvoyant research at clairvoyant abilities meaning.
“Early profits and published reports gave the field a data-driven reason to expand research.”
The Core Mechanics of Associative Remote Viewing for Predicting Future Events
A simple split-image design anchors each trial, with a facilitator pairing two sealed pictures to distinct outcomes. This step creates a blind mapping between a visual target and a specific result.
During the session, the percipient sketches or describes impressions without access to the mapping. These notes act as raw data that later get compared to the sealed targets.

Once the event unfolds, the facilitator reveals which image matched the actual outcome. That feedback closes the loop and provides a direct measure of match quality.
Researchers then score each trial by comparing the sketch to both targets. Over many trials, this process can produce statistically robust predictions of the outcome.
The protocol enforces strict blinding so the percipient cannot influence results. Repeated, controlled trials build confidence in any signal that appears above chance.
“Blinded pairing and consistent feedback are central to turning impressions into testable predictions.”
| Step | Action | Purpose |
|---|---|---|
| Selection | Facilitator pairs two sealed images | Create a blind link to outcomes |
| Session | Percipient sketches without mapping | Collect unbiased impressions |
| Reveal | Image shown after event | Provide feedback and scoring |
For a concise review of related research and abilities, see a short guide to clairvoyant abilities and research.
Standard Protocols Versus Modified Approaches
Comparing standard procedures to custom tweaks revealed how protocol design changed match rates and bias in many trials. Teams relied on clear steps to keep sessions testable and repeatable.
Standard Protocol Steps
The classic arv protocol, as used by Targ and Puthoff, had a facilitator select two sealed photos and assign them to binary outcomes. A percipient described impressions blind to that mapping. After the outcome occurred, the facilitator revealed which photo matched the actual result and researchers scored the match.

Modifications for Randomization
Researchers like Greg Kolodziejzyk introduced changes to increase randomization and reduce intentional influence. Some projects used shuffled digital libraries to pick target images. Others let the percipient self-judge to speed sessions and cut personnel bias.
- Large photo banks kept targets unpredictable.
- Self-judging combined facilitation and perception roles.
- High trial counts โ sometimes hundreds โ improved statistical power.
| Aspect | Standard | Modified |
|---|---|---|
| Image selection | Facilitator chosen photos | Randomized digital photo bank |
| Judging | Independent judges | Self-judging or automated scoring |
| Trial number | Dozens to hundreds | Tens to thousands depending on study |
“Strong blinding and varied photos raised confidence in any signal above chance.”
The Role of Consensus in Improving Accuracy
Aggregating many independent reports often reveals patterns that single trials miss.
Majority voting pools multiple sessions to create a single, stronger prediction. This approach reduces random noise and highlights repeated elements across notes or sketches.
Research support: Carpenter (1991) found that majority voting raised reliability when several percipients worked on the same task. The method complements an arv protocol by adding a statistical layer to human impressions.

The Power of Majority Voting
By running many trials and comparing sketches to each candidate photo, teams compute a consensus score. That score measures how often a particular image best matches across the number of sessions.
Consensus scoring improves the overall hit rate of predictions in published trials. It helps a remote viewer filter fleeting impressions and focus on consistent information.
| Metric | How It’s Measured | Benefit |
|---|---|---|
| Number of trials | Count of independent sessions | Increases statistical power |
| Consensus score | Aggregate match ratings per photo | Reduces single-trial error |
| Outcome accuracy | Percent correct matches | Improves with majority voting |
“Pooling simple judgments often turns scattered impressions into reliable signals.”
Designing a Robust Experimental Framework
A solid experimental plan starts with a wide, randomized photo bank that keeps targets unpredictable.
Maintain a large database of photos so the percipient never becomes familiar with images over repeated trials. This reduces learning effects and preserves test integrity.
Set the number of trials to fit statistical needs. Too few trials yield weak signals; too many can introduce fatigue. Plan a clear trial schedule and rest breaks.
Enforce strict blinding between facilitator and participant. Seal mappings, log times, and never allow hints that could leak information.

Use software to randomize photo selection and record the selection history. Automation makes the randomization transparent and repeatable across the project.
Document each trial with precise timing, scoring rules, and versioned data files. Independent reviewers should be able to validate the score and replicate the experiment from those records.
“Transparent randomization and careful timing turn anecdote into verifiable research.”
Selecting Future Events for Prediction
Choose events with clear, measurable outcomes so each trial yields an unambiguous score. That clarity helps teams record a clean match between a sketch and a photo.
Pick simple, binary targets โ coin flips, option expiry direction, or any outcome with two distinct states. These choices make scoring straightforward and boost statistical power across trials.
Researchers often use a random number generator to pick the specific event from a vetted basket. This step removes selection bias and keeps the project defensible.

Manage the time gap between session and event carefully. Too short a time can add noise. Too long can introduce unrelated changes. Set a fixed time window to keep each trial comparable.
Focus matters: when a percipient zeroes in on one target image tied to the outcome, their notes become easier to score. Clear definitions, stable timing, and random selection together raise the quality of arv predictions.
“Well-defined events and tight trial controls make matches easier to validate.”
Managing Target Image Databases
A well-managed image library is the backbone of any high-quality trial workflow. It ensures each trial uses a unique target and preserves the integrity of the project.

Digital Image Library Management
Keep a large number of high-resolution photos so targets never repeat inside a single project. This reduces learning effects and keeps matches unbiased.
Organize the database with clear tags and a fast search index. Quick retrieval saves setup time when scheduling each trial and reduces human error.
Automated random selection helps maintain blindness. Use vetted scripts or software to pull a single photo per trial and log the mapping securely.
- Scale: thousands of photos prevent reuse across a long project.
- Security: store mapping information encrypted until the trial ends.
- Audit: keep timestamps and logs so score reviews can verify each selection.
“A transparent, automated image system lets researchers run many trials with confidence.”
The Remote Viewing Session Process
A well-run session starts with quiet and small rituals that help the viewer settle into a focused state. The percipient sits in a low-noise room and often wears ear protectors to block distractions.
Many teams add soft tones like binaural beats to reach a theta brain wave state. This state is thought to help access subtle impressions and make sketches clearer.
Each trial asks the viewer to imagine a sealed photo as a single target image. They sketch impressions on a notepad and add brief notes about shapes, colors, and textures.
Sessions repeat multiple times per day depending on the number of trials the project needs. Keeping the schedule steady helps keep scores consistent and reduces fatigue.
At a set time the facilitator links the sketch to an outcome and reveals the matching photo. That reveal provides the information needed to score each trial and refine arv methods.

For practical guidance on session habits and interpretation, see a short guide to psychic intuitive readings.
Subjective Confidence Scoring and Its Impact
A short self-rated score after each session helps teams separate strong signals from weak noise. The viewer gives a numeric confidence level that rates how closely their sketch matches the sealed photo.
High scores matter: a score of 4 flags a trial as very reliable. Researchers treat that mark as evidence the impression is likely nonrandom and worth special attention.
Teams use these ratings to weight predictions in the project. Higher confidence trials carry more influence when aggregating results across many trials and photos.
Why this helps: by analyzing each trial’s score, analysts identify which viewing session entries repeatedly match the chosen target and outcome. That makes the dataset easier to act on, especially in financial applications.
- Scores filter low-quality data and tighten overall accuracy.
- Weighted tallies boost the signal from strong sessions in final predictions.
- Tracking score trends over time shows which viewers or protocols perform best.

“Confidence scores turn subjective impressions into a practical weighting system that improves prediction clarity.”
Analyzing Results from Long Term Studies
Large trial counts let analysts separate chance noise from a real signal.
Statistical Significance
Greg Kolodziejzyk ran a 13-year project and published the dataset in the Journal of Parapsychology. The study totaled 5,677 trials and compared sketches to a sealed photo paired with each target.
The headline result was a 52.65% success rate. That figure is statistically significant and sits above the 50% expected chance in a binary setup.
- The number of trials reduced random variation and raised confidence in the result.
- Statistical tests showed the effect persisted after controlling for obvious biases.

Long Term Data Trends
Long-term trends show that weighting trials by self-rated score lifts the success rate. When high-confidence trials are emphasized, aggregate rates can rise above 70%.
Effect size across many trials confirms a measurable effect. These long-term results support the view that anomalous cognition can produce reliable information when strict protocols and many trials are used.
“Large datasets turn scattered impressions into testable results.”
For practical applications and related methods, see a short guide to clairvoyant predictions.
Applying ARV to Financial Market Forecasting
Financial teams have adapted arv protocols to predict whether an asset will rise or fall over a defined time window.
In practice, a facilitator pairs two distinct photos with up/down market outcomes. A viewer sketches and rates impressions across several trials per day to build a consensus prediction.

Traders run many short sessions so the daily consensus reduces random noise and highlights repeated signals. Teams then use that aggregated prediction to set clear entry and exit rules on the market.
Project results from past experiments show this method can be practical when strict blinding and consistent scoring are enforced. Some groups reported profitable runs in futures trading after weighting high-score trials.
- Design each project question so the market outcome is binary and unambiguous.
- Run multiple trials per day to form a consensus prediction.
- Weight results by score to emphasize stronger impressions.
“When protocol, photos, and timing are strict, arv-based predictions can guide actionable market moves.”
Success Rates with Untrained Participants
Researchers ran a compact experiment that tested whether novices could use an arv protocol to forecast market direction. Smith et al. recruited untrained subjects and ran short, repeated trials tied to Dow Jones outcomes.
The group matched sketches to two sealed photos mapped to up or down market moves. Remarkably, participants hit a 7/7 success rate. That result was statistically significant and far above the expected chance.

These findings show the protocol can be used by people without formal background. The study argues that a simple, structured experiment can yield reliable predictions in a short project.
Multiple follow-up studies echoed similar patterns: untrained viewers often provide useful data when sessions are well controlled and feedback is regular.
“A clear protocol and many trials let even novice participants produce above-chance results.”
- Accessible protocol: novices can follow steps and give usable sketches.
- High hit rate: results exceeded chance in tightly controlled trials.
- Research implication: the effect suggests a latent human capacity worth further study.
Challenges in Maintaining Experimental Blinding
Keeping test conditions airtight is vital when investigators try to isolate subtle cognitive signals. Any information leak can undermine an entire experiment and skew the reported results.
Researchers must ensure the percipient stays fully blind to target images until the linked result is confirmed. Even small lapses can reduce the observed effect size and lower the hit rate in later studies.
Automation helps. Well-designed software randomizes image selection and logs mappings securely. That reduces human error and prevents accidental hints from facilitators.
Consistent protocol adherence is the final safeguard. Teams that follow a fixed arv protocol, keep timestamps, and enforce sealed mappings produce cleaner predictions and more defensible research outcomes.
“Maintaining strict blinding turns fragile impressions into verifiable results.”

| Risk | Likely Impact | Mitigation |
|---|---|---|
| Facilitator cueing | Bias in match ratings; lower effect | Automated randomization and audit logs |
| Image reuse | Familiarity reduces blindness | Large encrypted image bank |
| Poor timing | Variable scoring and noisy results | Fixed windows and strict timestamps |
For a concise primer on related extrasensory methods, see extra-sensory perception.
Future Directions for Parapsychology Research
A renewed focus on standard protocol design can make trials more comparable and easier to replicate. Researchers should prioritize a shared checklist for a clean remote viewing session to cut variability across labs.
Work will also dig deeper into the mechanisms behind anomalous cognition. Understanding how impressions arise could lift the overall effect seen in many studies.

Teams aim to increase effect size by refining ARV protocols and by testing new ways to engage the percipient. Better training, consistent timing, and automated randomization can improve the success rate.
The Journal of Parapsychology will remain a key outlet to publish rigorous results and peer reviews. Clear reporting and open data will let other groups replicate and extend promising findings.
“Standardization, transparent methods, and technology will help turn scattered signals into stronger, testable effects.”
- Standardize the viewing session setup and scoring rules.
- Investigate underlying mechanisms to explain observed effects.
- Use automation and larger datasets to raise the rate of reliable results.
Conclusion
Decades of controlled trials have produced consistent signals that outpace what we expect by chance. These results show a clear pattern: impressions often match the actual event more than the expected chance baseline.
The combined data, including the 13-year project by greg kolodziejzyk, strengthen the case that associative remote viewing can raise the success rate of a simple binary prediction. Small, well-run studies and a larger remote viewing experiment both delivered repeatable results in many trials.
Overall, the viewing experiment record and published work in the journal parapsychology suggest the method yields usable results about a future event. Continued study will refine protocols and improve the prediction rate.