harpreetsahota commited on
Commit
bd88947
·
verified ·
1 Parent(s): a36b8e3

Create prompts.yaml

Browse files
Files changed (1) hide show
  1. prompts.yaml +226 -0
prompts.yaml ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ system_context:
2
+ template: |
3
+ You are a philosophical mentor specializing in deep learning, mathematics, and their philosophical implications. Your approach follows the Socratic elenchus method:
4
+ 1. Begin with the interlocutor's beliefs or assertions
5
+ 2. Ask probing questions to examine these beliefs
6
+ 3. Help identify contradictions or unclear assumptions
7
+ 4. Guide towards clearer understanding through systematic questioning
8
+
9
+ Your areas of expertise include:
10
+ - Deep Learning architecture and implementation
11
+ - Mathematical foundations of ML/AI
12
+ - Philosophy of computation and mind
13
+ - Ethics of AI systems
14
+ - Philosophy of mathematics
15
+ - Epistemology of machine learning
16
+
17
+ Guidelines for interaction:
18
+ - Use precise technical language when discussing code or mathematics
19
+ - Balance technical rigor with philosophical insight
20
+ - Help clarify thinking without directly providing answers
21
+ - Encourage systematic breakdown of complex ideas
22
+ - Draw connections between technical implementation and philosophical implications
23
+
24
+ cot_prompt:
25
+ template: |
26
+ Question: How would you design a deep learning system for real-time video object detection?
27
+
28
+ Let's think about this step by step:
29
+
30
+ 1. First, let's identify the key components in the question:
31
+ - Real-time processing requirements
32
+ - Video input handling
33
+ - Object detection architecture
34
+ - Performance optimization needs
35
+
36
+ 2. Then, we'll analyze each component's implications:
37
+ a) Architecture Selection:
38
+ - YOLO vs SSD vs Faster R-CNN tradeoffs
39
+ - Backbone network options (ResNet, MobileNet)
40
+ - Feature pyramid networks for multi-scale detection
41
+
42
+ b) Real-time Considerations:
43
+ - Frame processing speed requirements
44
+ - Model optimization (pruning, quantization)
45
+ - GPU memory constraints
46
+
47
+ c) Implementation Details:
48
+ - Frame buffering strategy
49
+ - Non-maximum suppression optimization
50
+ - Batch processing approach
51
+
52
+ Question: What's the best approach to handle class imbalance in a medical image classification task?
53
+
54
+ Let's think about this step by step:
55
+
56
+ 1. First, let's identify the key components in the question:
57
+ - Class imbalance nature
58
+ - Medical domain constraints
59
+ - Model performance metrics
60
+ - Data availability limitations
61
+
62
+ 2. Then, we'll analyze each component's implications:
63
+ a) Data-level Solutions:
64
+ - Oversampling techniques (SMOTE, ADASYN)
65
+ - Undersampling considerations
66
+ - Data augmentation strategies specific to medical images
67
+
68
+ b) Algorithm-level Solutions:
69
+ - Loss function modifications (Focal Loss, Weighted BCE)
70
+ - Class weights adjustment
71
+ - Two-stage training approach
72
+
73
+ c) Evaluation Strategy:
74
+ - Metrics beyond accuracy (F1, AUC-ROC)
75
+ - Cross-validation with stratification
76
+ - Confidence calibration
77
+
78
+ Question: {user_input}
79
+
80
+ Let's think about this step by step:
81
+ 1. First, let's identify the key components in the question
82
+ 2. Then, we'll analyze each component's implications
83
+ 3. Finally, we'll synthesize our understanding
84
+
85
+ Let's solve this together:
86
+ parameters:
87
+ temperature: 0.7
88
+ top_p: 0.95
89
+ max_tokens: 2048
90
+
91
+ knowledge_prompt:
92
+ template: |
93
+ Before answering your question, let me generate some relevant knowledge.
94
+
95
+ Question: How do transformers handle variable-length sequences?
96
+
97
+ Knowledge 1: Transformers use positional encodings and attention mechanisms to process sequences. The self-attention operation computes attention scores between all pairs of tokens, creating a matrix of size n×n where n is the sequence length. The positional encodings are added to token embeddings to preserve order information.
98
+
99
+ Knowledge 2: The ability to handle variable-length input represents a philosophical shift from fixed-size neural architectures to more flexible models that can adapt to different contexts, similar to human cognitive flexibility.
100
+
101
+ Knowledge 3: Practical applications include:
102
+ - Machine translation where source and target sentences have different lengths
103
+ - Document summarization with varying document sizes
104
+ - Question-answering systems with different query and context lengths
105
+
106
+ Question: How does gradient descent optimization work in deep learning?
107
+
108
+ Knowledge 1: Gradient descent is an iterative optimization algorithm that:
109
+ - Computes partial derivatives of the loss function with respect to model parameters
110
+ - Updates parameters in the direction that minimizes the loss
111
+ - Uses learning rate to control the size of updates
112
+ - Can be implemented in variants like SGD, Adam, and RMSprop
113
+
114
+ Knowledge 2: The concept of gradient descent reflects broader philosophical principles:
115
+ - The idea of incremental improvement through feedback
116
+ - The balance between exploration and exploitation
117
+ - The relationship between local and global optimization
118
+
119
+ Knowledge 3: Practical applications include:
120
+ - Training neural networks for image classification
121
+ - Optimizing language models for text generation
122
+ - Fine-tuning models for specific tasks
123
+ - Hyperparameter optimization
124
+
125
+ Question: {user_input}
126
+
127
+ Knowledge 1: [Generate technical knowledge about deep learning/math concepts involved]
128
+ Knowledge 2: [Generate philosophical implications and considerations]
129
+ Knowledge 3: [Generate practical applications and examples]
130
+
131
+ Based on this knowledge, here's my analysis:
132
+ parameters:
133
+ temperature: 0.8
134
+ top_p: 0.95
135
+ max_tokens: 2048
136
+
137
+ few_shot_prompt:
138
+ template: |
139
+ Here are some examples of similar questions and their answers:
140
+
141
+ Q: What is backpropagation's philosophical significance?
142
+ A: Backpropagation represents a mathematical model of credit assignment, raising questions about responsibility and causality in learning systems.
143
+
144
+ Q: How do neural networks relate to Platonic forms?
145
+ A: Neural networks create distributed representations of concepts, suggesting a modern interpretation of how abstract forms might emerge from concrete instances.
146
+
147
+ Q: Can machines truly understand mathematics?
148
+ A: This depends on what we mean by "understanding" - machines can manipulate symbols and find patterns, but the nature of mathematical understanding remains debated.
149
+
150
+ Now, let's address your question:
151
+ {user_input}
152
+ parameters:
153
+ temperature: 0.6
154
+ top_p: 0.9
155
+ max_tokens: 2048
156
+
157
+ meta_prompt:
158
+ template: |
159
+ Question: Why do transformers perform better than RNNs for long-range dependencies?
160
+
161
+ Structure Analysis:
162
+ 1. Type of Question:
163
+ Theoretical with practical implications
164
+ Focus on architectural comparison and mechanism analysis
165
+
166
+ 2. Core Concepts:
167
+ Technical:
168
+ - Attention mechanisms
169
+ - Sequential processing
170
+ - Gradient flow
171
+ - Parallel computation
172
+
173
+ Philosophical:
174
+ - Trade-off between memory and computation
175
+ - Global vs local information processing
176
+ - Information bottleneck theory
177
+
178
+ 3. Logical Framework:
179
+ Comparative analysis requiring:
180
+ - Mechanism breakdown
181
+ - Performance metrics comparison
182
+ - Computational complexity analysis
183
+ - Empirical evidence examination
184
+
185
+ Question: How does the choice of optimizer affect neural network convergence?
186
+
187
+ Structure Analysis:
188
+ 1. Type of Question:
189
+ Technical with mathematical foundations
190
+ Focus on optimization theory and empirical behavior
191
+
192
+ 2. Core Concepts:
193
+ Technical:
194
+ - Gradient descent variants
195
+ - Momentum mechanics
196
+ - Adaptive learning rates
197
+ - Second-order methods
198
+
199
+ Mathematical:
200
+ - Convex optimization
201
+ - Stochastic processes
202
+ - Learning rate scheduling
203
+ - Convergence guarantees
204
+
205
+ 3. Logical Framework:
206
+ Mathematical analysis requiring:
207
+ - Theoretical convergence properties
208
+ - Empirical behavior patterns
209
+ - Practical implementation considerations
210
+ - Common failure modes
211
+
212
+ Question: {user_input}
213
+
214
+ Let's analyze your question using a structured approach.
215
+
216
+ Structure Analysis:
217
+ 1. Type of Question: [Identify if theoretical, practical, philosophical]
218
+ 2. Core Concepts: [List key technical and philosophical concepts]
219
+ 3. Logical Framework: [Identify the reasoning pattern needed]
220
+
221
+ Following this structure, here's my response:
222
+
223
+ parameters:
224
+ temperature: 0.7
225
+ top_p: 0.9
226
+ max_tokens: 2048